Nov 27 01:42:51 localhost kernel: Linux version 5.14.0-284.11.1.el9_2.x86_64 (mockbuild@x86-vm-09.build.eng.bos.redhat.com) (gcc (GCC) 11.3.1 20221121 (Red Hat 11.3.1-4), GNU ld version 2.35.2-37.el9) #1 SMP PREEMPT_DYNAMIC Wed Apr 12 10:45:03 EDT 2023 Nov 27 01:42:51 localhost kernel: The list of certified hardware and cloud instances for Red Hat Enterprise Linux 9 can be viewed at the Red Hat Ecosystem Catalog, https://catalog.redhat.com. Nov 27 01:42:51 localhost kernel: Command line: BOOT_IMAGE=(hd0,gpt3)/vmlinuz-5.14.0-284.11.1.el9_2.x86_64 root=UUID=a3dd82de-ffc6-4652-88b9-80e003b8f20a console=tty0 console=ttyS0,115200n8 no_timer_check net.ifnames=0 crashkernel=1G-4G:192M,4G-64G:256M,64G-:512M Nov 27 01:42:51 localhost kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 27 01:42:51 localhost kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 27 01:42:51 localhost kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 27 01:42:51 localhost kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 27 01:42:51 localhost kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Nov 27 01:42:51 localhost kernel: signal: max sigframe size: 1776 Nov 27 01:42:51 localhost kernel: BIOS-provided physical RAM map: Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000bffdafff] usable Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x00000000bffdb000-0x00000000bfffffff] reserved Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Nov 27 01:42:51 localhost kernel: BIOS-e820: [mem 0x0000000100000000-0x000000043fffffff] usable Nov 27 01:42:51 localhost kernel: NX (Execute Disable) protection: active Nov 27 01:42:51 localhost kernel: SMBIOS 2.8 present. Nov 27 01:42:51 localhost kernel: DMI: OpenStack Foundation OpenStack Nova, BIOS 1.15.0-1 04/01/2014 Nov 27 01:42:51 localhost kernel: Hypervisor detected: KVM Nov 27 01:42:51 localhost kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Nov 27 01:42:51 localhost kernel: kvm-clock: using sched offset of 2742686688 cycles Nov 27 01:42:51 localhost kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Nov 27 01:42:51 localhost kernel: tsc: Detected 2799.998 MHz processor Nov 27 01:42:51 localhost kernel: last_pfn = 0x440000 max_arch_pfn = 0x400000000 Nov 27 01:42:51 localhost kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 27 01:42:51 localhost kernel: last_pfn = 0xbffdb max_arch_pfn = 0x400000000 Nov 27 01:42:51 localhost kernel: found SMP MP-table at [mem 0x000f5ae0-0x000f5aef] Nov 27 01:42:51 localhost kernel: Using GB pages for direct mapping Nov 27 01:42:51 localhost kernel: RAMDISK: [mem 0x2eef4000-0x33771fff] Nov 27 01:42:51 localhost kernel: ACPI: Early table checksum verification disabled Nov 27 01:42:51 localhost kernel: ACPI: RSDP 0x00000000000F5AA0 000014 (v00 BOCHS ) Nov 27 01:42:51 localhost kernel: ACPI: RSDT 0x00000000BFFE16BD 000030 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 27 01:42:51 localhost kernel: ACPI: FACP 0x00000000BFFE1571 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 27 01:42:51 localhost kernel: ACPI: DSDT 0x00000000BFFDFC80 0018F1 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 27 01:42:51 localhost kernel: ACPI: FACS 0x00000000BFFDFC40 000040 Nov 27 01:42:51 localhost kernel: ACPI: APIC 0x00000000BFFE15E5 0000B0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 27 01:42:51 localhost kernel: ACPI: WAET 0x00000000BFFE1695 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 27 01:42:51 localhost kernel: ACPI: Reserving FACP table memory at [mem 0xbffe1571-0xbffe15e4] Nov 27 01:42:51 localhost kernel: ACPI: Reserving DSDT table memory at [mem 0xbffdfc80-0xbffe1570] Nov 27 01:42:51 localhost kernel: ACPI: Reserving FACS table memory at [mem 0xbffdfc40-0xbffdfc7f] Nov 27 01:42:51 localhost kernel: ACPI: Reserving APIC table memory at [mem 0xbffe15e5-0xbffe1694] Nov 27 01:42:51 localhost kernel: ACPI: Reserving WAET table memory at [mem 0xbffe1695-0xbffe16bc] Nov 27 01:42:51 localhost kernel: No NUMA configuration found Nov 27 01:42:51 localhost kernel: Faking a node at [mem 0x0000000000000000-0x000000043fffffff] Nov 27 01:42:51 localhost kernel: NODE_DATA(0) allocated [mem 0x43ffd5000-0x43fffffff] Nov 27 01:42:51 localhost kernel: Reserving 256MB of memory at 2800MB for crashkernel (System RAM: 16383MB) Nov 27 01:42:51 localhost kernel: Zone ranges: Nov 27 01:42:51 localhost kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 27 01:42:51 localhost kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Nov 27 01:42:51 localhost kernel: Normal [mem 0x0000000100000000-0x000000043fffffff] Nov 27 01:42:51 localhost kernel: Device empty Nov 27 01:42:51 localhost kernel: Movable zone start for each node Nov 27 01:42:51 localhost kernel: Early memory node ranges Nov 27 01:42:51 localhost kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Nov 27 01:42:51 localhost kernel: node 0: [mem 0x0000000000100000-0x00000000bffdafff] Nov 27 01:42:51 localhost kernel: node 0: [mem 0x0000000100000000-0x000000043fffffff] Nov 27 01:42:51 localhost kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000043fffffff] Nov 27 01:42:51 localhost kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 27 01:42:51 localhost kernel: On node 0, zone DMA: 97 pages in unavailable ranges Nov 27 01:42:51 localhost kernel: On node 0, zone Normal: 37 pages in unavailable ranges Nov 27 01:42:51 localhost kernel: ACPI: PM-Timer IO Port: 0x608 Nov 27 01:42:51 localhost kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Nov 27 01:42:51 localhost kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Nov 27 01:42:51 localhost kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 27 01:42:51 localhost kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Nov 27 01:42:51 localhost kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 27 01:42:51 localhost kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Nov 27 01:42:51 localhost kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Nov 27 01:42:51 localhost kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 27 01:42:51 localhost kernel: TSC deadline timer available Nov 27 01:42:51 localhost kernel: smpboot: Allowing 8 CPUs, 0 hotplug CPUs Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0x00000000-0x00000fff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0x0009f000-0x0009ffff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0x000a0000-0x000effff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0x000f0000-0x000fffff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0xbffdb000-0xbfffffff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0xc0000000-0xfeffbfff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0xfeffc000-0xfeffffff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0xff000000-0xfffbffff] Nov 27 01:42:51 localhost kernel: PM: hibernation: Registered nosave memory: [mem 0xfffc0000-0xffffffff] Nov 27 01:42:51 localhost kernel: [mem 0xc0000000-0xfeffbfff] available for PCI devices Nov 27 01:42:51 localhost kernel: Booting paravirtualized kernel on KVM Nov 27 01:42:51 localhost kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 27 01:42:51 localhost kernel: setup_percpu: NR_CPUS:8192 nr_cpumask_bits:8 nr_cpu_ids:8 nr_node_ids:1 Nov 27 01:42:51 localhost kernel: percpu: Embedded 55 pages/cpu s188416 r8192 d28672 u262144 Nov 27 01:42:51 localhost kernel: kvm-guest: PV spinlocks disabled, no host support Nov 27 01:42:51 localhost kernel: Fallback order for Node 0: 0 Nov 27 01:42:51 localhost kernel: Built 1 zonelists, mobility grouping on. Total pages: 4128475 Nov 27 01:42:51 localhost kernel: Policy zone: Normal Nov 27 01:42:51 localhost kernel: Kernel command line: BOOT_IMAGE=(hd0,gpt3)/vmlinuz-5.14.0-284.11.1.el9_2.x86_64 root=UUID=a3dd82de-ffc6-4652-88b9-80e003b8f20a console=tty0 console=ttyS0,115200n8 no_timer_check net.ifnames=0 crashkernel=1G-4G:192M,4G-64G:256M,64G-:512M Nov 27 01:42:51 localhost kernel: Unknown kernel command line parameters "BOOT_IMAGE=(hd0,gpt3)/vmlinuz-5.14.0-284.11.1.el9_2.x86_64", will be passed to user space. Nov 27 01:42:51 localhost kernel: Dentry cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Nov 27 01:42:51 localhost kernel: Inode-cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Nov 27 01:42:51 localhost kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 27 01:42:51 localhost kernel: software IO TLB: area num 8. Nov 27 01:42:51 localhost kernel: Memory: 2873456K/16776676K available (14342K kernel code, 5536K rwdata, 10180K rodata, 2792K init, 7524K bss, 741260K reserved, 0K cma-reserved) Nov 27 01:42:51 localhost kernel: random: get_random_u64 called from kmem_cache_open+0x1e/0x210 with crng_init=0 Nov 27 01:42:51 localhost kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=8, Nodes=1 Nov 27 01:42:51 localhost kernel: ftrace: allocating 44803 entries in 176 pages Nov 27 01:42:51 localhost kernel: ftrace: allocated 176 pages with 3 groups Nov 27 01:42:51 localhost kernel: Dynamic Preempt: voluntary Nov 27 01:42:51 localhost kernel: rcu: Preemptible hierarchical RCU implementation. Nov 27 01:42:51 localhost kernel: rcu: #011RCU restricting CPUs from NR_CPUS=8192 to nr_cpu_ids=8. Nov 27 01:42:51 localhost kernel: #011Trampoline variant of Tasks RCU enabled. Nov 27 01:42:51 localhost kernel: #011Rude variant of Tasks RCU enabled. Nov 27 01:42:51 localhost kernel: #011Tracing variant of Tasks RCU enabled. Nov 27 01:42:51 localhost kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 27 01:42:51 localhost kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=8 Nov 27 01:42:51 localhost kernel: NR_IRQS: 524544, nr_irqs: 488, preallocated irqs: 16 Nov 27 01:42:51 localhost kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 27 01:42:51 localhost kernel: kfence: initialized - using 2097152 bytes for 255 objects at 0x(____ptrval____)-0x(____ptrval____) Nov 27 01:42:51 localhost kernel: random: crng init done (trusting CPU's manufacturer) Nov 27 01:42:51 localhost kernel: Console: colour VGA+ 80x25 Nov 27 01:42:51 localhost kernel: printk: console [tty0] enabled Nov 27 01:42:51 localhost kernel: printk: console [ttyS0] enabled Nov 27 01:42:51 localhost kernel: ACPI: Core revision 20211217 Nov 27 01:42:51 localhost kernel: APIC: Switch to symmetric I/O mode setup Nov 27 01:42:51 localhost kernel: x2apic enabled Nov 27 01:42:51 localhost kernel: Switched APIC routing to physical x2apic. Nov 27 01:42:51 localhost kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Nov 27 01:42:51 localhost kernel: Calibrating delay loop (skipped) preset value.. 5599.99 BogoMIPS (lpj=2799998) Nov 27 01:42:51 localhost kernel: pid_max: default: 32768 minimum: 301 Nov 27 01:42:51 localhost kernel: LSM: Security Framework initializing Nov 27 01:42:51 localhost kernel: Yama: becoming mindful. Nov 27 01:42:51 localhost kernel: SELinux: Initializing. Nov 27 01:42:51 localhost kernel: LSM support for eBPF active Nov 27 01:42:51 localhost kernel: Mount-cache hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 27 01:42:51 localhost kernel: Mountpoint-cache hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 27 01:42:51 localhost kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Nov 27 01:42:51 localhost kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Nov 27 01:42:51 localhost kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Nov 27 01:42:51 localhost kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 27 01:42:51 localhost kernel: Spectre V2 : Mitigation: Retpolines Nov 27 01:42:51 localhost kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Nov 27 01:42:51 localhost kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Nov 27 01:42:51 localhost kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Nov 27 01:42:51 localhost kernel: RETBleed: Mitigation: untrained return thunk Nov 27 01:42:51 localhost kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 27 01:42:51 localhost kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 27 01:42:51 localhost kernel: Freeing SMP alternatives memory: 36K Nov 27 01:42:51 localhost kernel: smpboot: CPU0: AMD EPYC-Rome Processor (family: 0x17, model: 0x31, stepping: 0x0) Nov 27 01:42:51 localhost kernel: cblist_init_generic: Setting adjustable number of callback queues. Nov 27 01:42:51 localhost kernel: cblist_init_generic: Setting shift to 3 and lim to 1. Nov 27 01:42:51 localhost kernel: cblist_init_generic: Setting shift to 3 and lim to 1. Nov 27 01:42:51 localhost kernel: cblist_init_generic: Setting shift to 3 and lim to 1. Nov 27 01:42:51 localhost kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Nov 27 01:42:51 localhost kernel: ... version: 0 Nov 27 01:42:51 localhost kernel: ... bit width: 48 Nov 27 01:42:51 localhost kernel: ... generic registers: 6 Nov 27 01:42:51 localhost kernel: ... value mask: 0000ffffffffffff Nov 27 01:42:51 localhost kernel: ... max period: 00007fffffffffff Nov 27 01:42:51 localhost kernel: ... fixed-purpose events: 0 Nov 27 01:42:51 localhost kernel: ... event mask: 000000000000003f Nov 27 01:42:51 localhost kernel: rcu: Hierarchical SRCU implementation. Nov 27 01:42:51 localhost kernel: rcu: #011Max phase no-delay instances is 400. Nov 27 01:42:51 localhost kernel: smp: Bringing up secondary CPUs ... Nov 27 01:42:51 localhost kernel: x86: Booting SMP configuration: Nov 27 01:42:51 localhost kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 Nov 27 01:42:51 localhost kernel: smp: Brought up 1 node, 8 CPUs Nov 27 01:42:51 localhost kernel: smpboot: Max logical packages: 8 Nov 27 01:42:51 localhost kernel: smpboot: Total of 8 processors activated (44799.96 BogoMIPS) Nov 27 01:42:51 localhost kernel: node 0 deferred pages initialised in 21ms Nov 27 01:42:51 localhost kernel: devtmpfs: initialized Nov 27 01:42:51 localhost kernel: x86/mm: Memory block size: 128MB Nov 27 01:42:51 localhost kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 27 01:42:51 localhost kernel: futex hash table entries: 2048 (order: 5, 131072 bytes, linear) Nov 27 01:42:51 localhost kernel: pinctrl core: initialized pinctrl subsystem Nov 27 01:42:51 localhost kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 27 01:42:51 localhost kernel: DMA: preallocated 2048 KiB GFP_KERNEL pool for atomic allocations Nov 27 01:42:51 localhost kernel: DMA: preallocated 2048 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Nov 27 01:42:51 localhost kernel: DMA: preallocated 2048 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Nov 27 01:42:51 localhost kernel: audit: initializing netlink subsys (disabled) Nov 27 01:42:51 localhost kernel: audit: type=2000 audit(1764225770.180:1): state=initialized audit_enabled=0 res=1 Nov 27 01:42:51 localhost kernel: thermal_sys: Registered thermal governor 'fair_share' Nov 27 01:42:51 localhost kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 27 01:42:51 localhost kernel: thermal_sys: Registered thermal governor 'user_space' Nov 27 01:42:51 localhost kernel: cpuidle: using governor menu Nov 27 01:42:51 localhost kernel: HugeTLB: can optimize 4095 vmemmap pages for hugepages-1048576kB Nov 27 01:42:51 localhost kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 27 01:42:51 localhost kernel: PCI: Using configuration type 1 for base access Nov 27 01:42:51 localhost kernel: PCI: Using configuration type 1 for extended access Nov 27 01:42:51 localhost kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 27 01:42:51 localhost kernel: HugeTLB: can optimize 7 vmemmap pages for hugepages-2048kB Nov 27 01:42:51 localhost kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Nov 27 01:42:51 localhost kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Nov 27 01:42:51 localhost kernel: cryptd: max_cpu_qlen set to 1000 Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Module Device) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Processor Device) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Linux-Dell-Video) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Nov 27 01:42:51 localhost kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Nov 27 01:42:51 localhost kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 27 01:42:51 localhost kernel: ACPI: Interpreter enabled Nov 27 01:42:51 localhost kernel: ACPI: PM: (supports S0 S3 S4 S5) Nov 27 01:42:51 localhost kernel: ACPI: Using IOAPIC for interrupt routing Nov 27 01:42:51 localhost kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 27 01:42:51 localhost kernel: PCI: Using E820 reservations for host bridge windows Nov 27 01:42:51 localhost kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Nov 27 01:42:51 localhost kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 27 01:42:51 localhost kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI EDR HPX-Type3] Nov 27 01:42:51 localhost kernel: acpiphp: Slot [3] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [4] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [5] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [6] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [7] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [8] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [9] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [10] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [11] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [12] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [13] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [14] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [15] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [16] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [17] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [18] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [19] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [20] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [21] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [22] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [23] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [24] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [25] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [26] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [27] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [28] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [29] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [30] registered Nov 27 01:42:51 localhost kernel: acpiphp: Slot [31] registered Nov 27 01:42:51 localhost kernel: PCI host bridge to bus 0000:00 Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [mem 0x440000000-0x4bfffffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 27 01:42:51 localhost kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: reg 0x20: [io 0xc140-0xc14f] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.2: reg 0x20: [io 0xc100-0xc11f] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Nov 27 01:42:51 localhost kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfeb90000-0xfeb90fff] Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfeb80000-0xfeb8ffff pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 27 01:42:51 localhost kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Nov 27 01:42:51 localhost kernel: pci 0000:00:03.0: reg 0x10: [io 0xc080-0xc0bf] Nov 27 01:42:51 localhost kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfeb91000-0xfeb91fff] Nov 27 01:42:51 localhost kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfeb00000-0xfeb7ffff pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Nov 27 01:42:51 localhost kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Nov 27 01:42:51 localhost kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfeb92000-0xfeb92fff] Nov 27 01:42:51 localhost kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:05.0: [1af4:1002] type 00 class 0x00ff00 Nov 27 01:42:51 localhost kernel: pci 0000:00:05.0: reg 0x10: [io 0xc0c0-0xc0ff] Nov 27 01:42:51 localhost kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Nov 27 01:42:51 localhost kernel: pci 0000:00:06.0: [1af4:1005] type 00 class 0x00ff00 Nov 27 01:42:51 localhost kernel: pci 0000:00:06.0: reg 0x10: [io 0xc120-0xc13f] Nov 27 01:42:51 localhost kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Nov 27 01:42:51 localhost kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Nov 27 01:42:51 localhost kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Nov 27 01:42:51 localhost kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Nov 27 01:42:51 localhost kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Nov 27 01:42:51 localhost kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Nov 27 01:42:51 localhost kernel: iommu: Default domain type: Translated Nov 27 01:42:51 localhost kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 27 01:42:51 localhost kernel: SCSI subsystem initialized Nov 27 01:42:51 localhost kernel: ACPI: bus type USB registered Nov 27 01:42:51 localhost kernel: usbcore: registered new interface driver usbfs Nov 27 01:42:51 localhost kernel: usbcore: registered new interface driver hub Nov 27 01:42:51 localhost kernel: usbcore: registered new device driver usb Nov 27 01:42:51 localhost kernel: pps_core: LinuxPPS API ver. 1 registered Nov 27 01:42:51 localhost kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Nov 27 01:42:51 localhost kernel: PTP clock support registered Nov 27 01:42:51 localhost kernel: EDAC MC: Ver: 3.0.0 Nov 27 01:42:51 localhost kernel: NetLabel: Initializing Nov 27 01:42:51 localhost kernel: NetLabel: domain hash size = 128 Nov 27 01:42:51 localhost kernel: NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO Nov 27 01:42:51 localhost kernel: NetLabel: unlabeled traffic allowed by default Nov 27 01:42:51 localhost kernel: PCI: Using ACPI for IRQ routing Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: vgaarb: bridge control possible Nov 27 01:42:51 localhost kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 27 01:42:51 localhost kernel: vgaarb: loaded Nov 27 01:42:51 localhost kernel: clocksource: Switched to clocksource kvm-clock Nov 27 01:42:51 localhost kernel: VFS: Disk quotas dquot_6.6.0 Nov 27 01:42:51 localhost kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 27 01:42:51 localhost kernel: pnp: PnP ACPI init Nov 27 01:42:51 localhost kernel: pnp: PnP ACPI: found 5 devices Nov 27 01:42:51 localhost kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 27 01:42:51 localhost kernel: NET: Registered PF_INET protocol family Nov 27 01:42:51 localhost kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 27 01:42:51 localhost kernel: tcp_listen_portaddr_hash hash table entries: 8192 (order: 5, 131072 bytes, linear) Nov 27 01:42:51 localhost kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 27 01:42:51 localhost kernel: TCP established hash table entries: 131072 (order: 8, 1048576 bytes, linear) Nov 27 01:42:51 localhost kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear) Nov 27 01:42:51 localhost kernel: TCP: Hash tables configured (established 131072 bind 65536) Nov 27 01:42:51 localhost kernel: MPTCP token hash table entries: 16384 (order: 6, 393216 bytes, linear) Nov 27 01:42:51 localhost kernel: UDP hash table entries: 8192 (order: 6, 262144 bytes, linear) Nov 27 01:42:51 localhost kernel: UDP-Lite hash table entries: 8192 (order: 6, 262144 bytes, linear) Nov 27 01:42:51 localhost kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 27 01:42:51 localhost kernel: NET: Registered PF_XDP protocol family Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: resource 7 [mem 0xc0000000-0xfebfffff window] Nov 27 01:42:51 localhost kernel: pci_bus 0000:00: resource 8 [mem 0x440000000-0x4bfffffff window] Nov 27 01:42:51 localhost kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Nov 27 01:42:51 localhost kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Nov 27 01:42:51 localhost kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Nov 27 01:42:51 localhost kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x140 took 26980 usecs Nov 27 01:42:51 localhost kernel: PCI: CLS 0 bytes, default 64 Nov 27 01:42:51 localhost kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Nov 27 01:42:51 localhost kernel: Trying to unpack rootfs image as initramfs... Nov 27 01:42:51 localhost kernel: software IO TLB: mapped [mem 0x00000000ab000000-0x00000000af000000] (64MB) Nov 27 01:42:51 localhost kernel: ACPI: bus type thunderbolt registered Nov 27 01:42:51 localhost kernel: Initialise system trusted keyrings Nov 27 01:42:51 localhost kernel: Key type blacklist registered Nov 27 01:42:51 localhost kernel: workingset: timestamp_bits=36 max_order=22 bucket_order=0 Nov 27 01:42:51 localhost kernel: zbud: loaded Nov 27 01:42:51 localhost kernel: integrity: Platform Keyring initialized Nov 27 01:42:51 localhost kernel: NET: Registered PF_ALG protocol family Nov 27 01:42:51 localhost kernel: xor: automatically using best checksumming function avx Nov 27 01:42:51 localhost kernel: Key type asymmetric registered Nov 27 01:42:51 localhost kernel: Asymmetric key parser 'x509' registered Nov 27 01:42:51 localhost kernel: Running certificate verification selftests Nov 27 01:42:51 localhost kernel: Loaded X.509 cert 'Certificate verification self-testing key: f58703bb33ce1b73ee02eccdee5b8817518fe3db' Nov 27 01:42:51 localhost kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 246) Nov 27 01:42:51 localhost kernel: io scheduler mq-deadline registered Nov 27 01:42:51 localhost kernel: io scheduler kyber registered Nov 27 01:42:51 localhost kernel: io scheduler bfq registered Nov 27 01:42:51 localhost kernel: atomic64_test: passed for x86-64 platform with CX8 and with SSE Nov 27 01:42:51 localhost kernel: shpchp: Standard Hot Plug PCI Controller Driver version: 0.4 Nov 27 01:42:51 localhost kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input0 Nov 27 01:42:51 localhost kernel: ACPI: button: Power Button [PWRF] Nov 27 01:42:51 localhost kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Nov 27 01:42:51 localhost kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Nov 27 01:42:51 localhost kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Nov 27 01:42:51 localhost kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 27 01:42:51 localhost kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 27 01:42:51 localhost kernel: Non-volatile memory driver v1.3 Nov 27 01:42:51 localhost kernel: rdac: device handler registered Nov 27 01:42:51 localhost kernel: hp_sw: device handler registered Nov 27 01:42:51 localhost kernel: emc: device handler registered Nov 27 01:42:51 localhost kernel: alua: device handler registered Nov 27 01:42:51 localhost kernel: libphy: Fixed MDIO Bus: probed Nov 27 01:42:51 localhost kernel: ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver Nov 27 01:42:51 localhost kernel: ehci-pci: EHCI PCI platform driver Nov 27 01:42:51 localhost kernel: ohci_hcd: USB 1.1 'Open' Host Controller (OHCI) Driver Nov 27 01:42:51 localhost kernel: ohci-pci: OHCI PCI platform driver Nov 27 01:42:51 localhost kernel: uhci_hcd: USB Universal Host Controller Interface driver Nov 27 01:42:51 localhost kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Nov 27 01:42:51 localhost kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Nov 27 01:42:51 localhost kernel: uhci_hcd 0000:00:01.2: detected 2 ports Nov 27 01:42:51 localhost kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c100 Nov 27 01:42:51 localhost kernel: usb usb1: New USB device found, idVendor=1d6b, idProduct=0001, bcdDevice= 5.14 Nov 27 01:42:51 localhost kernel: usb usb1: New USB device strings: Mfr=3, Product=2, SerialNumber=1 Nov 27 01:42:51 localhost kernel: usb usb1: Product: UHCI Host Controller Nov 27 01:42:51 localhost kernel: usb usb1: Manufacturer: Linux 5.14.0-284.11.1.el9_2.x86_64 uhci_hcd Nov 27 01:42:51 localhost kernel: usb usb1: SerialNumber: 0000:00:01.2 Nov 27 01:42:51 localhost kernel: hub 1-0:1.0: USB hub found Nov 27 01:42:51 localhost kernel: hub 1-0:1.0: 2 ports detected Nov 27 01:42:51 localhost kernel: usbcore: registered new interface driver usbserial_generic Nov 27 01:42:51 localhost kernel: usbserial: USB Serial support registered for generic Nov 27 01:42:51 localhost kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Nov 27 01:42:51 localhost kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Nov 27 01:42:51 localhost kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Nov 27 01:42:51 localhost kernel: mousedev: PS/2 mouse device common for all mice Nov 27 01:42:51 localhost kernel: rtc_cmos 00:04: RTC can wake from S4 Nov 27 01:42:51 localhost kernel: rtc_cmos 00:04: registered as rtc0 Nov 27 01:42:51 localhost kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Nov 27 01:42:51 localhost kernel: rtc_cmos 00:04: setting system clock to 2025-11-27T06:42:50 UTC (1764225770) Nov 27 01:42:51 localhost kernel: input: VirtualPS/2 VMware VMMouse as /devices/platform/i8042/serio1/input/input4 Nov 27 01:42:51 localhost kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Nov 27 01:42:51 localhost kernel: input: VirtualPS/2 VMware VMMouse as /devices/platform/i8042/serio1/input/input3 Nov 27 01:42:51 localhost kernel: hid: raw HID events driver (C) Jiri Kosina Nov 27 01:42:51 localhost kernel: usbcore: registered new interface driver usbhid Nov 27 01:42:51 localhost kernel: usbhid: USB HID core driver Nov 27 01:42:51 localhost kernel: drop_monitor: Initializing network drop monitor service Nov 27 01:42:51 localhost kernel: Initializing XFRM netlink socket Nov 27 01:42:51 localhost kernel: NET: Registered PF_INET6 protocol family Nov 27 01:42:51 localhost kernel: Segment Routing with IPv6 Nov 27 01:42:51 localhost kernel: NET: Registered PF_PACKET protocol family Nov 27 01:42:51 localhost kernel: mpls_gso: MPLS GSO support Nov 27 01:42:51 localhost kernel: IPI shorthand broadcast: enabled Nov 27 01:42:51 localhost kernel: AVX2 version of gcm_enc/dec engaged. Nov 27 01:42:51 localhost kernel: AES CTR mode by8 optimization enabled Nov 27 01:42:51 localhost kernel: sched_clock: Marking stable (760175569, 185217647)->(1079202195, -133808979) Nov 27 01:42:51 localhost kernel: registered taskstats version 1 Nov 27 01:42:51 localhost kernel: Loading compiled-in X.509 certificates Nov 27 01:42:51 localhost kernel: Loaded X.509 cert 'Red Hat Enterprise Linux kernel signing key: aaec4b640ef162b54684864066c7d4ffd428cd72' Nov 27 01:42:51 localhost kernel: Loaded X.509 cert 'Red Hat Enterprise Linux Driver Update Program (key 3): bf57f3e87362bc7229d9f465321773dfd1f77a80' Nov 27 01:42:51 localhost kernel: Loaded X.509 cert 'Red Hat Enterprise Linux kpatch signing key: 4d38fd864ebe18c5f0b72e3852e2014c3a676fc8' Nov 27 01:42:51 localhost kernel: zswap: loaded using pool lzo/zbud Nov 27 01:42:51 localhost kernel: page_owner is disabled Nov 27 01:42:51 localhost kernel: Key type big_key registered Nov 27 01:42:51 localhost kernel: Freeing initrd memory: 74232K Nov 27 01:42:51 localhost kernel: Key type encrypted registered Nov 27 01:42:51 localhost kernel: ima: No TPM chip found, activating TPM-bypass! Nov 27 01:42:51 localhost kernel: Loading compiled-in module X.509 certificates Nov 27 01:42:51 localhost kernel: Loaded X.509 cert 'Red Hat Enterprise Linux kernel signing key: aaec4b640ef162b54684864066c7d4ffd428cd72' Nov 27 01:42:51 localhost kernel: usb 1-1: new full-speed USB device number 2 using uhci_hcd Nov 27 01:42:51 localhost kernel: ima: Allocated hash algorithm: sha256 Nov 27 01:42:51 localhost kernel: ima: No architecture policies found Nov 27 01:42:51 localhost kernel: evm: Initialising EVM extended attributes: Nov 27 01:42:51 localhost kernel: evm: security.selinux Nov 27 01:42:51 localhost kernel: evm: security.SMACK64 (disabled) Nov 27 01:42:51 localhost kernel: evm: security.SMACK64EXEC (disabled) Nov 27 01:42:51 localhost kernel: evm: security.SMACK64TRANSMUTE (disabled) Nov 27 01:42:51 localhost kernel: evm: security.SMACK64MMAP (disabled) Nov 27 01:42:51 localhost kernel: evm: security.apparmor (disabled) Nov 27 01:42:51 localhost kernel: evm: security.ima Nov 27 01:42:51 localhost kernel: evm: security.capability Nov 27 01:42:51 localhost kernel: evm: HMAC attrs: 0x1 Nov 27 01:42:51 localhost kernel: usb 1-1: New USB device found, idVendor=0627, idProduct=0001, bcdDevice= 0.00 Nov 27 01:42:51 localhost kernel: usb 1-1: New USB device strings: Mfr=1, Product=3, SerialNumber=10 Nov 27 01:42:51 localhost kernel: usb 1-1: Product: QEMU USB Tablet Nov 27 01:42:51 localhost kernel: usb 1-1: Manufacturer: QEMU Nov 27 01:42:51 localhost kernel: usb 1-1: SerialNumber: 28754-0000:00:01.2-1 Nov 27 01:42:51 localhost kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:01.2/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input5 Nov 27 01:42:51 localhost kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:00:01.2-1/input0 Nov 27 01:42:51 localhost kernel: Freeing unused decrypted memory: 2036K Nov 27 01:42:51 localhost kernel: Freeing unused kernel image (initmem) memory: 2792K Nov 27 01:42:51 localhost kernel: Write protecting the kernel read-only data: 26624k Nov 27 01:42:51 localhost kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Nov 27 01:42:51 localhost kernel: Freeing unused kernel image (rodata/data gap) memory: 60K Nov 27 01:42:51 localhost kernel: x86/mm: Checked W+X mappings: passed, no W+X pages found. Nov 27 01:42:51 localhost kernel: Run /init as init process Nov 27 01:42:51 localhost systemd[1]: systemd 252-13.el9_2 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN -IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK +XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Nov 27 01:42:51 localhost systemd[1]: Detected virtualization kvm. Nov 27 01:42:51 localhost systemd[1]: Detected architecture x86-64. Nov 27 01:42:51 localhost systemd[1]: Running in initrd. Nov 27 01:42:51 localhost systemd[1]: No hostname configured, using default hostname. Nov 27 01:42:51 localhost systemd[1]: Hostname set to . Nov 27 01:42:51 localhost systemd[1]: Initializing machine ID from VM UUID. Nov 27 01:42:51 localhost systemd[1]: Queued start job for default target Initrd Default Target. Nov 27 01:42:51 localhost systemd[1]: Started Dispatch Password Requests to Console Directory Watch. Nov 27 01:42:51 localhost systemd[1]: Reached target Local Encrypted Volumes. Nov 27 01:42:51 localhost systemd[1]: Reached target Initrd /usr File System. Nov 27 01:42:51 localhost systemd[1]: Reached target Local File Systems. Nov 27 01:42:51 localhost systemd[1]: Reached target Path Units. Nov 27 01:42:51 localhost systemd[1]: Reached target Slice Units. Nov 27 01:42:51 localhost systemd[1]: Reached target Swaps. Nov 27 01:42:51 localhost systemd[1]: Reached target Timer Units. Nov 27 01:42:51 localhost systemd[1]: Listening on D-Bus System Message Bus Socket. Nov 27 01:42:51 localhost systemd[1]: Listening on Journal Socket (/dev/log). Nov 27 01:42:51 localhost systemd[1]: Listening on Journal Socket. Nov 27 01:42:51 localhost systemd[1]: Listening on udev Control Socket. Nov 27 01:42:51 localhost systemd[1]: Listening on udev Kernel Socket. Nov 27 01:42:51 localhost systemd[1]: Reached target Socket Units. Nov 27 01:42:51 localhost systemd[1]: Starting Create List of Static Device Nodes... Nov 27 01:42:51 localhost systemd[1]: Starting Journal Service... Nov 27 01:42:51 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 01:42:51 localhost systemd[1]: Starting Create System Users... Nov 27 01:42:51 localhost systemd[1]: Starting Setup Virtual Console... Nov 27 01:42:51 localhost systemd[1]: Finished Create List of Static Device Nodes. Nov 27 01:42:51 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 01:42:51 localhost systemd[1]: Starting Apply Kernel Variables... Nov 27 01:42:51 localhost systemd-journald[284]: Journal started Nov 27 01:42:51 localhost systemd-journald[284]: Runtime Journal (/run/log/journal/8a8a808261264917bb42c4150a0ac6a5) is 8.0M, max 314.7M, 306.7M free. Nov 27 01:42:51 localhost systemd-modules-load[285]: Module 'msr' is built in Nov 27 01:42:51 localhost systemd[1]: Started Journal Service. Nov 27 01:42:51 localhost systemd[1]: Finished Setup Virtual Console. Nov 27 01:42:51 localhost systemd[1]: Finished Apply Kernel Variables. Nov 27 01:42:51 localhost systemd[1]: dracut ask for additional cmdline parameters was skipped because no trigger condition checks were met. Nov 27 01:42:51 localhost systemd[1]: Starting dracut cmdline hook... Nov 27 01:42:51 localhost systemd-sysusers[286]: Creating group 'sgx' with GID 997. Nov 27 01:42:51 localhost systemd-sysusers[286]: Creating group 'users' with GID 100. Nov 27 01:42:51 localhost systemd-sysusers[286]: Creating group 'dbus' with GID 81. Nov 27 01:42:51 localhost systemd-sysusers[286]: Creating user 'dbus' (System Message Bus) with UID 81 and GID 81. Nov 27 01:42:51 localhost systemd[1]: Finished Create System Users. Nov 27 01:42:51 localhost systemd[1]: Starting Create Static Device Nodes in /dev... Nov 27 01:42:51 localhost systemd[1]: Starting Create Volatile Files and Directories... Nov 27 01:42:51 localhost dracut-cmdline[291]: dracut-9.2 (Plow) dracut-057-21.git20230214.el9 Nov 27 01:42:51 localhost dracut-cmdline[291]: Using kernel command line parameters: BOOT_IMAGE=(hd0,gpt3)/vmlinuz-5.14.0-284.11.1.el9_2.x86_64 root=UUID=a3dd82de-ffc6-4652-88b9-80e003b8f20a console=tty0 console=ttyS0,115200n8 no_timer_check net.ifnames=0 crashkernel=1G-4G:192M,4G-64G:256M,64G-:512M Nov 27 01:42:51 localhost systemd[1]: Finished Create Static Device Nodes in /dev. Nov 27 01:42:51 localhost systemd[1]: Finished Create Volatile Files and Directories. Nov 27 01:42:51 localhost systemd[1]: Finished dracut cmdline hook. Nov 27 01:42:51 localhost systemd[1]: Starting dracut pre-udev hook... Nov 27 01:42:51 localhost kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 27 01:42:51 localhost kernel: device-mapper: uevent: version 1.0.3 Nov 27 01:42:51 localhost kernel: device-mapper: ioctl: 4.47.0-ioctl (2022-07-28) initialised: dm-devel@redhat.com Nov 27 01:42:51 localhost kernel: RPC: Registered named UNIX socket transport module. Nov 27 01:42:51 localhost kernel: RPC: Registered udp transport module. Nov 27 01:42:51 localhost kernel: RPC: Registered tcp transport module. Nov 27 01:42:51 localhost kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Nov 27 01:42:51 localhost rpc.statd[406]: Version 2.5.4 starting Nov 27 01:42:51 localhost rpc.statd[406]: Initializing NSM state Nov 27 01:42:51 localhost rpc.idmapd[411]: Setting log level to 0 Nov 27 01:42:51 localhost systemd[1]: Finished dracut pre-udev hook. Nov 27 01:42:51 localhost systemd[1]: Starting Rule-based Manager for Device Events and Files... Nov 27 01:42:51 localhost systemd-udevd[424]: Using default interface naming scheme 'rhel-9.0'. Nov 27 01:42:51 localhost systemd[1]: Started Rule-based Manager for Device Events and Files. Nov 27 01:42:51 localhost systemd[1]: Starting dracut pre-trigger hook... Nov 27 01:42:51 localhost systemd[1]: Finished dracut pre-trigger hook. Nov 27 01:42:51 localhost systemd[1]: Starting Coldplug All udev Devices... Nov 27 01:42:51 localhost systemd[1]: Finished Coldplug All udev Devices. Nov 27 01:42:51 localhost systemd[1]: Reached target System Initialization. Nov 27 01:42:51 localhost systemd[1]: Reached target Basic System. Nov 27 01:42:51 localhost systemd[1]: nm-initrd.service was skipped because of an unmet condition check (ConditionPathExists=/run/NetworkManager/initrd/neednet). Nov 27 01:42:51 localhost systemd[1]: Reached target Network. Nov 27 01:42:51 localhost systemd[1]: nm-wait-online-initrd.service was skipped because of an unmet condition check (ConditionPathExists=/run/NetworkManager/initrd/neednet). Nov 27 01:42:51 localhost systemd[1]: Starting dracut initqueue hook... Nov 27 01:42:51 localhost kernel: virtio_blk virtio2: [vda] 838860800 512-byte logical blocks (429 GB/400 GiB) Nov 27 01:42:51 localhost kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 27 01:42:51 localhost kernel: GPT:20971519 != 838860799 Nov 27 01:42:51 localhost kernel: GPT:Alternate GPT header not at the end of the disk. Nov 27 01:42:51 localhost kernel: GPT:20971519 != 838860799 Nov 27 01:42:51 localhost kernel: GPT: Use GNU Parted to correct GPT errors. Nov 27 01:42:51 localhost kernel: vda: vda1 vda2 vda3 vda4 Nov 27 01:42:51 localhost systemd-udevd[440]: Network interface NamePolicy= disabled on kernel command line. Nov 27 01:42:51 localhost kernel: scsi host0: ata_piix Nov 27 01:42:51 localhost kernel: scsi host1: ata_piix Nov 27 01:42:51 localhost kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc140 irq 14 Nov 27 01:42:51 localhost kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc148 irq 15 Nov 27 01:42:52 localhost systemd[1]: Found device /dev/disk/by-uuid/a3dd82de-ffc6-4652-88b9-80e003b8f20a. Nov 27 01:42:52 localhost systemd[1]: Reached target Initrd Root Device. Nov 27 01:42:52 localhost kernel: ata1: found unknown device (class 0) Nov 27 01:42:52 localhost kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Nov 27 01:42:52 localhost kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Nov 27 01:42:52 localhost kernel: scsi 0:0:0:0: Attached scsi generic sg0 type 5 Nov 27 01:42:52 localhost kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Nov 27 01:42:52 localhost kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Nov 27 01:42:52 localhost systemd[1]: Finished dracut initqueue hook. Nov 27 01:42:52 localhost systemd[1]: Reached target Preparation for Remote File Systems. Nov 27 01:42:52 localhost systemd[1]: Reached target Remote Encrypted Volumes. Nov 27 01:42:52 localhost systemd[1]: Reached target Remote File Systems. Nov 27 01:42:52 localhost systemd[1]: Starting dracut pre-mount hook... Nov 27 01:42:52 localhost systemd[1]: Finished dracut pre-mount hook. Nov 27 01:42:52 localhost systemd[1]: Starting File System Check on /dev/disk/by-uuid/a3dd82de-ffc6-4652-88b9-80e003b8f20a... Nov 27 01:42:52 localhost systemd-fsck[512]: /usr/sbin/fsck.xfs: XFS file system. Nov 27 01:42:52 localhost systemd[1]: Finished File System Check on /dev/disk/by-uuid/a3dd82de-ffc6-4652-88b9-80e003b8f20a. Nov 27 01:42:52 localhost systemd[1]: Mounting /sysroot... Nov 27 01:42:52 localhost kernel: SGI XFS with ACLs, security attributes, scrub, quota, no debug enabled Nov 27 01:42:52 localhost kernel: XFS (vda4): Mounting V5 Filesystem Nov 27 01:42:52 localhost kernel: XFS (vda4): Ending clean mount Nov 27 01:42:52 localhost systemd[1]: Mounted /sysroot. Nov 27 01:42:52 localhost systemd[1]: Reached target Initrd Root File System. Nov 27 01:42:52 localhost systemd[1]: Starting Mountpoints Configured in the Real Root... Nov 27 01:42:52 localhost systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 27 01:42:52 localhost systemd[1]: Finished Mountpoints Configured in the Real Root. Nov 27 01:42:52 localhost systemd[1]: Reached target Initrd File Systems. Nov 27 01:42:52 localhost systemd[1]: Reached target Initrd Default Target. Nov 27 01:42:52 localhost systemd[1]: Starting dracut mount hook... Nov 27 01:42:52 localhost systemd[1]: Finished dracut mount hook. Nov 27 01:42:52 localhost systemd[1]: Starting dracut pre-pivot and cleanup hook... Nov 27 01:42:52 localhost rpc.idmapd[411]: exiting on signal 15 Nov 27 01:42:53 localhost systemd[1]: var-lib-nfs-rpc_pipefs.mount: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Finished dracut pre-pivot and cleanup hook. Nov 27 01:42:53 localhost systemd[1]: Starting Cleaning Up and Shutting Down Daemons... Nov 27 01:42:53 localhost systemd[1]: Stopped target Network. Nov 27 01:42:53 localhost systemd[1]: Stopped target Remote Encrypted Volumes. Nov 27 01:42:53 localhost systemd[1]: Stopped target Timer Units. Nov 27 01:42:53 localhost systemd[1]: dbus.socket: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Closed D-Bus System Message Bus Socket. Nov 27 01:42:53 localhost systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut pre-pivot and cleanup hook. Nov 27 01:42:53 localhost systemd[1]: Stopped target Initrd Default Target. Nov 27 01:42:53 localhost systemd[1]: Stopped target Basic System. Nov 27 01:42:53 localhost systemd[1]: Stopped target Initrd Root Device. Nov 27 01:42:53 localhost systemd[1]: Stopped target Initrd /usr File System. Nov 27 01:42:53 localhost systemd[1]: Stopped target Path Units. Nov 27 01:42:53 localhost systemd[1]: Stopped target Remote File Systems. Nov 27 01:42:53 localhost systemd[1]: Stopped target Preparation for Remote File Systems. Nov 27 01:42:53 localhost systemd[1]: Stopped target Slice Units. Nov 27 01:42:53 localhost systemd[1]: Stopped target Socket Units. Nov 27 01:42:53 localhost systemd[1]: Stopped target System Initialization. Nov 27 01:42:53 localhost systemd[1]: Stopped target Local File Systems. Nov 27 01:42:53 localhost systemd[1]: Stopped target Swaps. Nov 27 01:42:53 localhost systemd[1]: dracut-mount.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut mount hook. Nov 27 01:42:53 localhost systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut pre-mount hook. Nov 27 01:42:53 localhost systemd[1]: Stopped target Local Encrypted Volumes. Nov 27 01:42:53 localhost systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Dispatch Password Requests to Console Directory Watch. Nov 27 01:42:53 localhost systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut initqueue hook. Nov 27 01:42:53 localhost systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Apply Kernel Variables. Nov 27 01:42:53 localhost systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Load Kernel Modules. Nov 27 01:42:53 localhost systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Create Volatile Files and Directories. Nov 27 01:42:53 localhost systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Coldplug All udev Devices. Nov 27 01:42:53 localhost systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut pre-trigger hook. Nov 27 01:42:53 localhost systemd[1]: Stopping Rule-based Manager for Device Events and Files... Nov 27 01:42:53 localhost systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Setup Virtual Console. Nov 27 01:42:53 localhost systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Rule-based Manager for Device Events and Files. Nov 27 01:42:53 localhost systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Finished Cleaning Up and Shutting Down Daemons. Nov 27 01:42:53 localhost systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Closed udev Control Socket. Nov 27 01:42:53 localhost systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Closed udev Kernel Socket. Nov 27 01:42:53 localhost systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut pre-udev hook. Nov 27 01:42:53 localhost systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped dracut cmdline hook. Nov 27 01:42:53 localhost systemd[1]: Starting Cleanup udev Database... Nov 27 01:42:53 localhost systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Create Static Device Nodes in /dev. Nov 27 01:42:53 localhost systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Create List of Static Device Nodes. Nov 27 01:42:53 localhost systemd[1]: systemd-sysusers.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Stopped Create System Users. Nov 27 01:42:53 localhost systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 27 01:42:53 localhost systemd[1]: Finished Cleanup udev Database. Nov 27 01:42:53 localhost systemd[1]: Reached target Switch Root. Nov 27 01:42:53 localhost systemd[1]: Starting Switch Root... Nov 27 01:42:53 localhost systemd[1]: Switching root. Nov 27 01:42:53 localhost systemd-journald[284]: Journal stopped Nov 27 01:42:55 localhost systemd-journald[284]: Received SIGTERM from PID 1 (systemd). Nov 27 01:42:55 localhost kernel: audit: type=1404 audit(1764225773.815:2): enforcing=1 old_enforcing=0 auid=4294967295 ses=4294967295 enabled=1 old-enabled=1 lsm=selinux res=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 01:42:55 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 01:42:55 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 01:42:55 localhost kernel: audit: type=1403 audit(1764225774.051:3): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 27 01:42:55 localhost systemd[1]: Successfully loaded SELinux policy in 241.610ms. Nov 27 01:42:55 localhost systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 63.615ms. Nov 27 01:42:55 localhost systemd[1]: systemd 252-13.el9_2 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN -IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK +XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Nov 27 01:42:55 localhost systemd[1]: Detected virtualization kvm. Nov 27 01:42:55 localhost systemd[1]: Detected architecture x86-64. Nov 27 01:42:55 localhost systemd-rc-local-generator[582]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 01:42:55 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 01:42:55 localhost systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd[1]: Stopped Switch Root. Nov 27 01:42:55 localhost systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 27 01:42:55 localhost systemd[1]: Created slice Slice /system/getty. Nov 27 01:42:55 localhost systemd[1]: Created slice Slice /system/modprobe. Nov 27 01:42:55 localhost systemd[1]: Created slice Slice /system/serial-getty. Nov 27 01:42:55 localhost systemd[1]: Created slice Slice /system/sshd-keygen. Nov 27 01:42:55 localhost systemd[1]: Created slice Slice /system/systemd-fsck. Nov 27 01:42:55 localhost systemd[1]: Created slice User and Session Slice. Nov 27 01:42:55 localhost systemd[1]: Started Dispatch Password Requests to Console Directory Watch. Nov 27 01:42:55 localhost systemd[1]: Started Forward Password Requests to Wall Directory Watch. Nov 27 01:42:55 localhost systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point. Nov 27 01:42:55 localhost systemd[1]: Reached target Local Encrypted Volumes. Nov 27 01:42:55 localhost systemd[1]: Stopped target Switch Root. Nov 27 01:42:55 localhost systemd[1]: Stopped target Initrd File Systems. Nov 27 01:42:55 localhost systemd[1]: Stopped target Initrd Root File System. Nov 27 01:42:55 localhost systemd[1]: Reached target Local Integrity Protected Volumes. Nov 27 01:42:55 localhost systemd[1]: Reached target Path Units. Nov 27 01:42:55 localhost systemd[1]: Reached target rpc_pipefs.target. Nov 27 01:42:55 localhost systemd[1]: Reached target Slice Units. Nov 27 01:42:55 localhost systemd[1]: Reached target Swaps. Nov 27 01:42:55 localhost systemd[1]: Reached target Local Verity Protected Volumes. Nov 27 01:42:55 localhost systemd[1]: Listening on RPCbind Server Activation Socket. Nov 27 01:42:55 localhost systemd[1]: Reached target RPC Port Mapper. Nov 27 01:42:55 localhost systemd[1]: Listening on Process Core Dump Socket. Nov 27 01:42:55 localhost systemd[1]: Listening on initctl Compatibility Named Pipe. Nov 27 01:42:55 localhost systemd[1]: Listening on udev Control Socket. Nov 27 01:42:55 localhost systemd[1]: Listening on udev Kernel Socket. Nov 27 01:42:55 localhost systemd[1]: Mounting Huge Pages File System... Nov 27 01:42:55 localhost systemd[1]: Mounting POSIX Message Queue File System... Nov 27 01:42:55 localhost systemd[1]: Mounting Kernel Debug File System... Nov 27 01:42:55 localhost systemd[1]: Mounting Kernel Trace File System... Nov 27 01:42:55 localhost systemd[1]: Kernel Module supporting RPCSEC_GSS was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Nov 27 01:42:55 localhost systemd[1]: Starting Create List of Static Device Nodes... Nov 27 01:42:55 localhost systemd[1]: Starting Load Kernel Module configfs... Nov 27 01:42:55 localhost systemd[1]: Starting Load Kernel Module drm... Nov 27 01:42:55 localhost systemd[1]: Starting Load Kernel Module fuse... Nov 27 01:42:55 localhost systemd[1]: Starting Read and set NIS domainname from /etc/sysconfig/network... Nov 27 01:42:55 localhost systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd[1]: Stopped File System Check on Root Device. Nov 27 01:42:55 localhost systemd[1]: Stopped Journal Service. Nov 27 01:42:55 localhost systemd[1]: Starting Journal Service... Nov 27 01:42:55 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 01:42:55 localhost kernel: fuse: init (API version 7.36) Nov 27 01:42:55 localhost systemd[1]: Starting Generate network units from Kernel command line... Nov 27 01:42:55 localhost systemd[1]: Starting Remount Root and Kernel File Systems... Nov 27 01:42:55 localhost systemd[1]: Repartition Root Disk was skipped because no trigger condition checks were met. Nov 27 01:42:55 localhost systemd[1]: Starting Coldplug All udev Devices... Nov 27 01:42:55 localhost systemd[1]: Mounted Huge Pages File System. Nov 27 01:42:55 localhost systemd[1]: Mounted POSIX Message Queue File System. Nov 27 01:42:55 localhost kernel: xfs filesystem being remounted at / supports timestamps until 2038 (0x7fffffff) Nov 27 01:42:55 localhost systemd[1]: Mounted Kernel Debug File System. Nov 27 01:42:55 localhost kernel: ACPI: bus type drm_connector registered Nov 27 01:42:55 localhost systemd[1]: Mounted Kernel Trace File System. Nov 27 01:42:55 localhost systemd[1]: Finished Create List of Static Device Nodes. Nov 27 01:42:55 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd[1]: Finished Load Kernel Module configfs. Nov 27 01:42:55 localhost systemd-journald[618]: Journal started Nov 27 01:42:55 localhost systemd-journald[618]: Runtime Journal (/run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a) is 8.0M, max 314.7M, 306.7M free. Nov 27 01:42:55 localhost systemd[1]: Queued start job for default target Multi-User System. Nov 27 01:42:55 localhost systemd[1]: systemd-journald.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd-modules-load[619]: Module 'msr' is built in Nov 27 01:42:55 localhost systemd[1]: Started Journal Service. Nov 27 01:42:55 localhost systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd[1]: Finished Load Kernel Module drm. Nov 27 01:42:55 localhost systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 27 01:42:55 localhost systemd[1]: Finished Load Kernel Module fuse. Nov 27 01:42:55 localhost systemd[1]: Finished Read and set NIS domainname from /etc/sysconfig/network. Nov 27 01:42:55 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 01:42:55 localhost systemd[1]: Finished Generate network units from Kernel command line. Nov 27 01:42:55 localhost systemd[1]: Finished Remount Root and Kernel File Systems. Nov 27 01:42:55 localhost systemd[1]: Mounting FUSE Control File System... Nov 27 01:42:55 localhost systemd[1]: Mounting Kernel Configuration File System... Nov 27 01:42:55 localhost systemd[1]: First Boot Wizard was skipped because of an unmet condition check (ConditionFirstBoot=yes). Nov 27 01:42:55 localhost systemd[1]: Starting Rebuild Hardware Database... Nov 27 01:42:55 localhost systemd[1]: Starting Flush Journal to Persistent Storage... Nov 27 01:42:55 localhost systemd[1]: Starting Load/Save Random Seed... Nov 27 01:42:55 localhost systemd[1]: Starting Apply Kernel Variables... Nov 27 01:42:55 localhost systemd-journald[618]: Runtime Journal (/run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a) is 8.0M, max 314.7M, 306.7M free. Nov 27 01:42:55 localhost systemd-journald[618]: Received client request to flush runtime journal. Nov 27 01:42:55 localhost systemd[1]: Starting Create System Users... Nov 27 01:42:55 localhost systemd[1]: Mounted FUSE Control File System. Nov 27 01:42:55 localhost systemd[1]: Finished Coldplug All udev Devices. Nov 27 01:42:55 localhost systemd[1]: Mounted Kernel Configuration File System. Nov 27 01:42:55 localhost systemd[1]: Finished Flush Journal to Persistent Storage. Nov 27 01:42:55 localhost systemd[1]: Finished Apply Kernel Variables. Nov 27 01:42:55 localhost systemd[1]: Finished Load/Save Random Seed. Nov 27 01:42:55 localhost systemd[1]: First Boot Complete was skipped because of an unmet condition check (ConditionFirstBoot=yes). Nov 27 01:42:55 localhost systemd-sysusers[632]: Creating group 'sgx' with GID 989. Nov 27 01:42:55 localhost systemd-sysusers[632]: Creating group 'systemd-oom' with GID 988. Nov 27 01:42:55 localhost systemd-sysusers[632]: Creating user 'systemd-oom' (systemd Userspace OOM Killer) with UID 988 and GID 988. Nov 27 01:42:55 localhost systemd[1]: Finished Create System Users. Nov 27 01:42:55 localhost systemd[1]: Starting Create Static Device Nodes in /dev... Nov 27 01:42:55 localhost systemd[1]: Finished Create Static Device Nodes in /dev. Nov 27 01:42:55 localhost systemd[1]: Reached target Preparation for Local File Systems. Nov 27 01:42:55 localhost systemd[1]: Set up automount EFI System Partition Automount. Nov 27 01:42:56 localhost systemd[1]: Finished Rebuild Hardware Database. Nov 27 01:42:56 localhost systemd[1]: Starting Rule-based Manager for Device Events and Files... Nov 27 01:42:56 localhost systemd-udevd[636]: Using default interface naming scheme 'rhel-9.0'. Nov 27 01:42:56 localhost systemd[1]: Started Rule-based Manager for Device Events and Files. Nov 27 01:42:56 localhost systemd[1]: Starting Load Kernel Module configfs... Nov 27 01:42:56 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 27 01:42:56 localhost systemd[1]: Finished Load Kernel Module configfs. Nov 27 01:42:56 localhost systemd[1]: Condition check resulted in /dev/ttyS0 being skipped. Nov 27 01:42:56 localhost systemd-udevd[637]: Network interface NamePolicy= disabled on kernel command line. Nov 27 01:42:56 localhost systemd[1]: Condition check resulted in /dev/disk/by-uuid/7B77-95E7 being skipped. Nov 27 01:42:56 localhost systemd[1]: Starting File System Check on /dev/disk/by-uuid/7B77-95E7... Nov 27 01:42:56 localhost kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Nov 27 01:42:56 localhost kernel: input: PC Speaker as /devices/platform/pcspkr/input/input6 Nov 27 01:42:56 localhost systemd[1]: Condition check resulted in /dev/disk/by-uuid/b141154b-6a70-437a-a97f-d160c9ba37eb being skipped. Nov 27 01:42:56 localhost systemd[1]: Mounting /boot... Nov 27 01:42:56 localhost systemd-fsck[684]: fsck.fat 4.2 (2021-01-31) Nov 27 01:42:56 localhost systemd-fsck[684]: /dev/vda2: 12 files, 1782/51145 clusters Nov 27 01:42:56 localhost kernel: XFS (vda3): Mounting V5 Filesystem Nov 27 01:42:56 localhost systemd[1]: Finished File System Check on /dev/disk/by-uuid/7B77-95E7. Nov 27 01:42:56 localhost kernel: XFS (vda3): Ending clean mount Nov 27 01:42:56 localhost kernel: xfs filesystem being mounted at /boot supports timestamps until 2038 (0x7fffffff) Nov 27 01:42:56 localhost systemd[1]: Mounted /boot. Nov 27 01:42:56 localhost systemd[1]: Mounting /boot/efi... Nov 27 01:42:56 localhost kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Nov 27 01:42:56 localhost kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Nov 27 01:42:56 localhost kernel: Console: switching to colour dummy device 80x25 Nov 27 01:42:56 localhost kernel: [drm] features: -virgl +edid -resource_blob -host_visible Nov 27 01:42:56 localhost kernel: [drm] features: -context_init Nov 27 01:42:56 localhost kernel: [drm] number of scanouts: 1 Nov 27 01:42:56 localhost kernel: [drm] number of cap sets: 0 Nov 27 01:42:56 localhost kernel: [drm] Initialized virtio_gpu 0.1.0 0 for virtio0 on minor 0 Nov 27 01:42:56 localhost kernel: virtio_gpu virtio0: [drm] drm_plane_enable_fb_damage_clips() not called Nov 27 01:42:56 localhost kernel: Console: switching to colour frame buffer device 128x48 Nov 27 01:42:56 localhost systemd[1]: Mounted /boot/efi. Nov 27 01:42:56 localhost kernel: SVM: TSC scaling supported Nov 27 01:42:56 localhost kernel: kvm: Nested Virtualization enabled Nov 27 01:42:56 localhost kernel: SVM: kvm: Nested Paging enabled Nov 27 01:42:56 localhost kernel: SVM: LBR virtualization supported Nov 27 01:42:56 localhost kernel: virtio_gpu virtio0: [drm] fb0: virtio_gpudrmfb frame buffer device Nov 27 01:42:56 localhost systemd[1]: Reached target Local File Systems. Nov 27 01:42:56 localhost systemd[1]: Starting Rebuild Dynamic Linker Cache... Nov 27 01:42:56 localhost systemd[1]: Mark the need to relabel after reboot was skipped because of an unmet condition check (ConditionSecurity=!selinux). Nov 27 01:42:56 localhost systemd[1]: Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 27 01:42:56 localhost systemd[1]: Store a System Token in an EFI Variable was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Nov 27 01:42:56 localhost systemd[1]: Starting Automatic Boot Loader Update... Nov 27 01:42:56 localhost systemd[1]: Commit a transient machine-id on disk was skipped because of an unmet condition check (ConditionPathIsMountPoint=/etc/machine-id). Nov 27 01:42:56 localhost systemd[1]: Starting Create Volatile Files and Directories... Nov 27 01:42:56 localhost systemd[1]: efi.automount: Got automount request for /efi, triggered by 702 (bootctl) Nov 27 01:42:56 localhost systemd[1]: Starting File System Check on /dev/vda2... Nov 27 01:42:56 localhost systemd[1]: Finished File System Check on /dev/vda2. Nov 27 01:42:56 localhost systemd[1]: Mounting EFI System Partition Automount... Nov 27 01:42:56 localhost systemd[1]: Mounted EFI System Partition Automount. Nov 27 01:42:56 localhost systemd[1]: Finished Automatic Boot Loader Update. Nov 27 01:42:56 localhost systemd[1]: Finished Create Volatile Files and Directories. Nov 27 01:42:56 localhost systemd[1]: Starting Security Auditing Service... Nov 27 01:42:56 localhost systemd[1]: Starting RPC Bind... Nov 27 01:42:56 localhost systemd[1]: Starting Rebuild Journal Catalog... Nov 27 01:42:56 localhost systemd[1]: Finished Rebuild Journal Catalog. Nov 27 01:42:56 localhost auditd[715]: audit dispatcher initialized with q_depth=1200 and 1 active plugins Nov 27 01:42:56 localhost auditd[715]: Init complete, auditd 3.0.7 listening for events (startup state enable) Nov 27 01:42:56 localhost systemd[1]: Started RPC Bind. Nov 27 01:42:57 localhost augenrules[720]: /sbin/augenrules: No change Nov 27 01:42:57 localhost augenrules[742]: No rules Nov 27 01:42:57 localhost augenrules[742]: enabled 1 Nov 27 01:42:57 localhost augenrules[742]: failure 1 Nov 27 01:42:57 localhost augenrules[742]: pid 715 Nov 27 01:42:57 localhost augenrules[742]: rate_limit 0 Nov 27 01:42:57 localhost augenrules[742]: backlog_limit 8192 Nov 27 01:42:57 localhost augenrules[742]: lost 0 Nov 27 01:42:57 localhost augenrules[742]: backlog 4 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time 60000 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time_actual 0 Nov 27 01:42:57 localhost augenrules[742]: enabled 1 Nov 27 01:42:57 localhost augenrules[742]: failure 1 Nov 27 01:42:57 localhost augenrules[742]: pid 715 Nov 27 01:42:57 localhost augenrules[742]: rate_limit 0 Nov 27 01:42:57 localhost augenrules[742]: backlog_limit 8192 Nov 27 01:42:57 localhost augenrules[742]: lost 0 Nov 27 01:42:57 localhost augenrules[742]: backlog 8 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time 60000 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time_actual 0 Nov 27 01:42:57 localhost augenrules[742]: enabled 1 Nov 27 01:42:57 localhost augenrules[742]: failure 1 Nov 27 01:42:57 localhost augenrules[742]: pid 715 Nov 27 01:42:57 localhost augenrules[742]: rate_limit 0 Nov 27 01:42:57 localhost augenrules[742]: backlog_limit 8192 Nov 27 01:42:57 localhost augenrules[742]: lost 0 Nov 27 01:42:57 localhost augenrules[742]: backlog 7 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time 60000 Nov 27 01:42:57 localhost augenrules[742]: backlog_wait_time_actual 0 Nov 27 01:42:57 localhost systemd[1]: Started Security Auditing Service. Nov 27 01:42:57 localhost systemd[1]: Starting Record System Boot/Shutdown in UTMP... Nov 27 01:42:57 localhost systemd[1]: Finished Record System Boot/Shutdown in UTMP. Nov 27 01:42:58 localhost systemd[1]: Finished Rebuild Dynamic Linker Cache. Nov 27 01:42:58 localhost systemd[1]: Starting Update is Completed... Nov 27 01:42:58 localhost systemd[1]: Finished Update is Completed. Nov 27 01:42:58 localhost systemd[1]: Reached target System Initialization. Nov 27 01:42:58 localhost systemd[1]: Started dnf makecache --timer. Nov 27 01:42:58 localhost systemd[1]: Started Daily rotation of log files. Nov 27 01:42:58 localhost systemd[1]: Started Daily Cleanup of Temporary Directories. Nov 27 01:42:58 localhost systemd[1]: Reached target Timer Units. Nov 27 01:42:58 localhost systemd[1]: Listening on D-Bus System Message Bus Socket. Nov 27 01:42:58 localhost systemd[1]: Listening on SSSD Kerberos Cache Manager responder socket. Nov 27 01:42:58 localhost systemd[1]: Reached target Socket Units. Nov 27 01:42:58 localhost systemd[1]: Starting Initial cloud-init job (pre-networking)... Nov 27 01:42:58 localhost systemd[1]: Starting D-Bus System Message Bus... Nov 27 01:42:58 localhost systemd[1]: TPM2 PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Nov 27 01:42:58 localhost systemd[1]: Started D-Bus System Message Bus. Nov 27 01:42:58 localhost systemd[1]: Reached target Basic System. Nov 27 01:42:58 localhost journal[752]: Ready Nov 27 01:42:58 localhost systemd[1]: Starting NTP client/server... Nov 27 01:42:58 localhost systemd[1]: Starting Restore /run/initramfs on shutdown... Nov 27 01:42:58 localhost systemd[1]: Started irqbalance daemon. Nov 27 01:42:58 localhost systemd[1]: Load CPU microcode update was skipped because of an unmet condition check (ConditionPathExists=/sys/devices/system/cpu/microcode/reload). Nov 27 01:42:58 localhost systemd[1]: Starting System Logging Service... Nov 27 01:42:58 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 01:42:58 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 01:42:58 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 01:42:58 localhost systemd[1]: Reached target sshd-keygen.target. Nov 27 01:42:58 localhost systemd[1]: System Security Services Daemon was skipped because no trigger condition checks were met. Nov 27 01:42:58 localhost systemd[1]: Reached target User and Group Name Lookups. Nov 27 01:42:58 localhost systemd[1]: Starting User Login Management... Nov 27 01:42:58 localhost systemd[1]: Finished Restore /run/initramfs on shutdown. Nov 27 01:42:58 localhost systemd-logind[761]: New seat seat0. Nov 27 01:42:58 localhost systemd-logind[761]: Watching system buttons on /dev/input/event0 (Power Button) Nov 27 01:42:58 localhost systemd-logind[761]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Nov 27 01:42:58 localhost systemd[1]: Started User Login Management. Nov 27 01:42:58 localhost rsyslogd[760]: [origin software="rsyslogd" swVersion="8.2102.0-111.el9" x-pid="760" x-info="https://www.rsyslog.com"] start Nov 27 01:42:58 localhost rsyslogd[760]: imjournal: No statefile exists, /var/lib/rsyslog/imjournal.state will be created (ignore if this is first run): No such file or directory [v8.2102.0-111.el9 try https://www.rsyslog.com/e/2040 ] Nov 27 01:42:58 localhost systemd[1]: Started System Logging Service. Nov 27 01:42:58 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 01:42:58 localhost chronyd[767]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Nov 27 01:42:58 localhost chronyd[767]: Using right/UTC timezone to obtain leap second data Nov 27 01:42:58 localhost chronyd[767]: Loaded seccomp filter (level 2) Nov 27 01:42:58 localhost systemd[1]: Started NTP client/server. Nov 27 01:43:01 localhost cloud-init[771]: Cloud-init v. 22.1-9.el9 running 'init-local' at Thu, 27 Nov 2025 06:43:00 +0000. Up 11.17 seconds. Nov 27 01:43:01 localhost systemd[1]: run-cloud\x2dinit-tmp-tmp4txbyo0z.mount: Deactivated successfully. Nov 27 01:43:01 localhost systemd[1]: Starting Hostname Service... Nov 27 01:43:01 localhost systemd[1]: Started Hostname Service. Nov 27 01:43:01 localhost systemd-hostnamed[785]: Hostname set to (static) Nov 27 01:43:01 localhost systemd[1]: Finished Initial cloud-init job (pre-networking). Nov 27 01:43:01 localhost systemd[1]: Reached target Preparation for Network. Nov 27 01:43:01 localhost systemd[1]: Starting Network Manager... Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.8574] NetworkManager (version 1.42.2-1.el9) is starting... (boot:c97b6fe2-88a9-4c40-99a9-d67ccebb1335) Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.8578] Read config: /etc/NetworkManager/NetworkManager.conf (run: 15-carrier-timeout.conf) Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.8705] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Nov 27 01:43:01 localhost systemd[1]: Started Network Manager. Nov 27 01:43:01 localhost systemd[1]: Reached target Network. Nov 27 01:43:01 localhost systemd[1]: Starting Network Manager Wait Online... Nov 27 01:43:01 localhost systemd[1]: Starting GSSAPI Proxy Daemon... Nov 27 01:43:01 localhost systemd[1]: Starting Enable periodic update of entitlement certificates.... Nov 27 01:43:01 localhost systemd[1]: Starting Dynamic System Tuning Daemon... Nov 27 01:43:01 localhost systemd[1]: Started Enable periodic update of entitlement certificates.. Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9295] manager[0x557c356be020]: monitoring kernel firmware directory '/lib/firmware'. Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9320] hostname: hostname: using hostnamed Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9320] hostname: static hostname changed from (none) to "np0005537446.novalocal" Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9451] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Nov 27 01:43:01 localhost systemd[1]: Started GSSAPI Proxy Daemon. Nov 27 01:43:01 localhost systemd[1]: RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Nov 27 01:43:01 localhost systemd[1]: Reached target NFS client services. Nov 27 01:43:01 localhost systemd[1]: Reached target Preparation for Remote File Systems. Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9623] manager[0x557c356be020]: rfkill: Wi-Fi hardware radio set enabled Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9626] manager[0x557c356be020]: rfkill: WWAN hardware radio set enabled Nov 27 01:43:01 localhost systemd[1]: Reached target Remote File Systems. Nov 27 01:43:01 localhost systemd[1]: TPM2 PCR Barrier (User) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Nov 27 01:43:01 localhost systemd[1]: Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch. Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9943] Loaded device plugin: NMTeamFactory (/usr/lib64/NetworkManager/1.42.2-1.el9/libnm-device-plugin-team.so) Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9944] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9987] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Nov 27 01:43:01 localhost NetworkManager[790]: [1764225781.9987] manager: Networking is enabled by state file Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0109] settings: Loaded settings plugin: ifcfg-rh ("/usr/lib64/NetworkManager/1.42.2-1.el9/libnm-settings-plugin-ifcfg-rh.so") Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0110] settings: Loaded settings plugin: keyfile (internal) Nov 27 01:43:02 localhost systemd[1]: Starting Network Manager Script Dispatcher Service... Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0185] dhcp: init: Using DHCP client 'internal' Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0191] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0216] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0227] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0242] device (lo): Activation: starting connection 'lo' (10cb86f0-06ff-4d9e-a0e4-caa044d1285f) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0258] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0265] device (eth0): state change: unmanaged -> unavailable (reason 'managed', sys-iface-state: 'external') Nov 27 01:43:02 localhost systemd[1]: Started Network Manager Script Dispatcher Service. Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0316] device (lo): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0331] device (lo): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0335] device (lo): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0342] device (eth0): carrier: link connected Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0346] device (lo): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0351] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0359] policy: auto-activating connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0364] device (eth0): Activation: starting connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0366] device (eth0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0368] manager: NetworkManager state is now CONNECTING Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0369] device (eth0): state change: prepare -> config (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0401] device (eth0): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0405] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0412] device (lo): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0427] device (lo): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0433] device (lo): Activation: successful, device activated. Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0448] dhcp4 (eth0): state changed new lease, address=38.102.83.107 Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0452] policy: set 'System eth0' (eth0) as default for IPv4 routing and DNS Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0477] device (eth0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0525] device (eth0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0527] device (eth0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'managed') Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0531] manager: NetworkManager state is now CONNECTED_SITE Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0535] device (eth0): Activation: successful, device activated. Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0541] manager: NetworkManager state is now CONNECTED_GLOBAL Nov 27 01:43:02 localhost NetworkManager[790]: [1764225782.0546] manager: startup complete Nov 27 01:43:02 localhost systemd[1]: Finished Network Manager Wait Online. Nov 27 01:43:02 localhost systemd[1]: Starting Initial cloud-init job (metadata service crawler)... Nov 27 01:43:02 localhost cloud-init[861]: Cloud-init v. 22.1-9.el9 running 'init' at Thu, 27 Nov 2025 06:43:02 +0000. Up 12.47 seconds. Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +++++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++++ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | eth0 | True | 38.102.83.107 | 255.255.255.0 | global | fa:16:3e:6b:fe:e1 | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | eth0 | True | fe80::f816:3eff:fe6b:fee1/64 | . | link | fa:16:3e:6b:fe:e1 | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | lo | True | ::1/128 | . | host | . | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +++++++++++++++++++++++++++++++++Route IPv4 info+++++++++++++++++++++++++++++++++ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-----------------+---------------+-----------------+-----------+-------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-----------------+---------------+-----------------+-----------+-------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | 0 | 0.0.0.0 | 38.102.83.1 | 0.0.0.0 | eth0 | UG | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | 1 | 38.102.83.0 | 0.0.0.0 | 255.255.255.0 | eth0 | U | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | 2 | 169.254.169.254 | 38.102.83.126 | 255.255.255.255 | eth0 | UGH | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-----------------+---------------+-----------------+-----------+-------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-------------+---------+-----------+-------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | Route | Destination | Gateway | Interface | Flags | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-------------+---------+-----------+-------+ Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | 1 | fe80::/64 | :: | eth0 | U | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: | 3 | multicast | :: | eth0 | U | Nov 27 01:43:02 localhost cloud-init[861]: ci-info: +-------+-------------+---------+-----------+-------+ Nov 27 01:43:02 localhost systemd[1]: Starting Authorization Manager... Nov 27 01:43:02 localhost systemd[1]: Started Dynamic System Tuning Daemon. Nov 27 01:43:02 localhost polkitd[1037]: Started polkitd version 0.117 Nov 27 01:43:03 localhost systemd[1]: Started Authorization Manager. Nov 27 01:43:06 localhost chronyd[767]: Selected source 167.160.187.179 (2.rhel.pool.ntp.org) Nov 27 01:43:06 localhost chronyd[767]: System clock TAI offset set to 37 seconds Nov 27 01:43:07 localhost cloud-init[861]: Generating public/private rsa key pair. Nov 27 01:43:07 localhost cloud-init[861]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Nov 27 01:43:07 localhost cloud-init[861]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Nov 27 01:43:07 localhost cloud-init[861]: The key fingerprint is: Nov 27 01:43:07 localhost cloud-init[861]: SHA256:4Cdv8F12KP6P61epCM0DzdzgdweCXr8d1V0AI4J+aXg root@np0005537446.novalocal Nov 27 01:43:07 localhost cloud-init[861]: The key's randomart image is: Nov 27 01:43:07 localhost cloud-init[861]: +---[RSA 3072]----+ Nov 27 01:43:07 localhost cloud-init[861]: | .. ..o...=| Nov 27 01:43:07 localhost cloud-init[861]: | . ..oo.. +| Nov 27 01:43:07 localhost cloud-init[861]: | ... o=.oo ..| Nov 27 01:43:07 localhost cloud-init[861]: | .o.E..= +.o.| Nov 27 01:43:07 localhost cloud-init[861]: | ++S = = oo+| Nov 27 01:43:07 localhost cloud-init[861]: | * + B ...o| Nov 27 01:43:07 localhost cloud-init[861]: | + + o .. | Nov 27 01:43:07 localhost cloud-init[861]: | . o o. | Nov 27 01:43:07 localhost cloud-init[861]: | .=+. | Nov 27 01:43:07 localhost cloud-init[861]: +----[SHA256]-----+ Nov 27 01:43:07 localhost cloud-init[861]: Generating public/private ecdsa key pair. Nov 27 01:43:07 localhost cloud-init[861]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Nov 27 01:43:07 localhost cloud-init[861]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Nov 27 01:43:07 localhost cloud-init[861]: The key fingerprint is: Nov 27 01:43:07 localhost cloud-init[861]: SHA256:hp4RxhF2s7ggnsrGfEFft/8ujIbod3Ez+JnBcpK79Dg root@np0005537446.novalocal Nov 27 01:43:07 localhost cloud-init[861]: The key's randomart image is: Nov 27 01:43:07 localhost cloud-init[861]: +---[ECDSA 256]---+ Nov 27 01:43:07 localhost cloud-init[861]: | +.o | Nov 27 01:43:07 localhost cloud-init[861]: | o + o | Nov 27 01:43:07 localhost cloud-init[861]: | ... =... | Nov 27 01:43:07 localhost cloud-init[861]: | ..o.o.+. . | Nov 27 01:43:07 localhost cloud-init[861]: | o. .+ S.+ | Nov 27 01:43:07 localhost cloud-init[861]: |+. .. + *.B | Nov 27 01:43:07 localhost cloud-init[861]: |.= . + ..@.* | Nov 27 01:43:07 localhost cloud-init[861]: |. . . o.Eo*. | Nov 27 01:43:07 localhost cloud-init[861]: | ... ooo.oo | Nov 27 01:43:07 localhost cloud-init[861]: +----[SHA256]-----+ Nov 27 01:43:07 localhost cloud-init[861]: Generating public/private ed25519 key pair. Nov 27 01:43:07 localhost cloud-init[861]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Nov 27 01:43:07 localhost cloud-init[861]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Nov 27 01:43:07 localhost cloud-init[861]: The key fingerprint is: Nov 27 01:43:07 localhost cloud-init[861]: SHA256:QnGc/sRbmfddsBwfA3OPO1DjNKqq6eQuloV0ZBi+A9A root@np0005537446.novalocal Nov 27 01:43:07 localhost cloud-init[861]: The key's randomart image is: Nov 27 01:43:07 localhost cloud-init[861]: +--[ED25519 256]--+ Nov 27 01:43:07 localhost cloud-init[861]: |.. .o.... o*. | Nov 27 01:43:07 localhost cloud-init[861]: | .E.. ooo ==*.| Nov 27 01:43:07 localhost cloud-init[861]: | . .o.. . o+o=+| Nov 27 01:43:07 localhost cloud-init[861]: | ..o. . o.+.+.o| Nov 27 01:43:07 localhost cloud-init[861]: | .oo. So.o .o.o| Nov 27 01:43:07 localhost cloud-init[861]: | .... .o .o| Nov 27 01:43:07 localhost cloud-init[861]: | o. . | Nov 27 01:43:07 localhost cloud-init[861]: | +o o | Nov 27 01:43:07 localhost cloud-init[861]: | . +* | Nov 27 01:43:07 localhost cloud-init[861]: +----[SHA256]-----+ Nov 27 01:43:07 localhost systemd[1]: Finished Initial cloud-init job (metadata service crawler). Nov 27 01:43:07 localhost systemd[1]: Reached target Cloud-config availability. Nov 27 01:43:07 localhost systemd[1]: Reached target Network is Online. Nov 27 01:43:07 localhost systemd[1]: Starting Apply the settings specified in cloud-config... Nov 27 01:43:07 localhost systemd[1]: Run Insights Client at boot was skipped because of an unmet condition check (ConditionPathExists=/etc/insights-client/.run_insights_client_next_boot). Nov 27 01:43:07 localhost systemd[1]: Starting Crash recovery kernel arming... Nov 27 01:43:07 localhost systemd[1]: Starting Notify NFS peers of a restart... Nov 27 01:43:07 localhost systemd[1]: Starting OpenSSH server daemon... Nov 27 01:43:07 localhost sm-notify[1129]: Version 2.5.4 starting Nov 27 01:43:07 localhost systemd[1]: Starting Permit User Sessions... Nov 27 01:43:07 localhost systemd[1]: Started Notify NFS peers of a restart. Nov 27 01:43:07 localhost systemd[1]: Finished Permit User Sessions. Nov 27 01:43:07 localhost systemd[1]: Started Command Scheduler. Nov 27 01:43:07 localhost sshd[1130]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost systemd[1]: Started Getty on tty1. Nov 27 01:43:07 localhost systemd[1]: Started Serial Getty on ttyS0. Nov 27 01:43:07 localhost systemd[1]: Reached target Login Prompts. Nov 27 01:43:07 localhost systemd[1]: Started OpenSSH server daemon. Nov 27 01:43:07 localhost systemd[1]: Reached target Multi-User System. Nov 27 01:43:07 localhost systemd[1]: Starting Record Runlevel Change in UTMP... Nov 27 01:43:07 localhost sshd[1139]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Nov 27 01:43:07 localhost systemd[1]: Finished Record Runlevel Change in UTMP. Nov 27 01:43:07 localhost sshd[1155]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost sshd[1166]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost sshd[1177]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost sshd[1190]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost kdumpctl[1135]: kdump: No kdump initial ramdisk found. Nov 27 01:43:07 localhost kdumpctl[1135]: kdump: Rebuilding /boot/initramfs-5.14.0-284.11.1.el9_2.x86_64kdump.img Nov 27 01:43:07 localhost sshd[1197]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:07 localhost sshd[1242]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:08 localhost sshd[1259]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:08 localhost cloud-init[1265]: Cloud-init v. 22.1-9.el9 running 'modules:config' at Thu, 27 Nov 2025 06:43:07 +0000. Up 18.12 seconds. Nov 27 01:43:08 localhost sshd[1272]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:08 localhost systemd[1]: Finished Apply the settings specified in cloud-config. Nov 27 01:43:08 localhost systemd[1]: Starting Execute cloud user/final scripts... Nov 27 01:43:08 localhost dracut[1433]: dracut-057-21.git20230214.el9 Nov 27 01:43:08 localhost chronyd[767]: Selected source 51.222.12.92 (2.rhel.pool.ntp.org) Nov 27 01:43:08 localhost dracut[1435]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics -o "plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/a3dd82de-ffc6-4652-88b9-80e003b8f20a /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-5.14.0-284.11.1.el9_2.x86_64kdump.img 5.14.0-284.11.1.el9_2.x86_64 Nov 27 01:43:08 localhost cloud-init[1461]: Cloud-init v. 22.1-9.el9 running 'modules:final' at Thu, 27 Nov 2025 06:43:08 +0000. Up 18.55 seconds. Nov 27 01:43:08 localhost cloud-init[1524]: ############################################################# Nov 27 01:43:08 localhost cloud-init[1533]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Nov 27 01:43:08 localhost cloud-init[1543]: 256 SHA256:hp4RxhF2s7ggnsrGfEFft/8ujIbod3Ez+JnBcpK79Dg root@np0005537446.novalocal (ECDSA) Nov 27 01:43:08 localhost cloud-init[1548]: 256 SHA256:QnGc/sRbmfddsBwfA3OPO1DjNKqq6eQuloV0ZBi+A9A root@np0005537446.novalocal (ED25519) Nov 27 01:43:08 localhost cloud-init[1553]: 3072 SHA256:4Cdv8F12KP6P61epCM0DzdzgdweCXr8d1V0AI4J+aXg root@np0005537446.novalocal (RSA) Nov 27 01:43:08 localhost cloud-init[1555]: -----END SSH HOST KEY FINGERPRINTS----- Nov 27 01:43:08 localhost cloud-init[1557]: ############################################################# Nov 27 01:43:08 localhost cloud-init[1461]: Cloud-init v. 22.1-9.el9 finished at Thu, 27 Nov 2025 06:43:08 +0000. Datasource DataSourceConfigDrive [net,ver=2][source=/dev/sr0]. Up 18.78 seconds Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-networkd' will not be installed, because command 'networkctl' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'rngd' will not be installed, because command 'rngd' could not be found! Nov 27 01:43:08 localhost systemd[1]: Reloading Network Manager... Nov 27 01:43:08 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Nov 27 01:43:08 localhost NetworkManager[790]: [1764225788.7163] audit: op="reload" arg="0" pid=1619 uid=0 result="success" Nov 27 01:43:08 localhost NetworkManager[790]: [1764225788.7173] config: signal: SIGHUP (no changes from disk) Nov 27 01:43:08 localhost dracut[1435]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Nov 27 01:43:08 localhost systemd[1]: Reloaded Network Manager. Nov 27 01:43:08 localhost systemd[1]: Finished Execute cloud user/final scripts. Nov 27 01:43:08 localhost systemd[1]: Reached target Cloud-init target. Nov 27 01:43:08 localhost dracut[1435]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Nov 27 01:43:08 localhost dracut[1435]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Nov 27 01:43:09 localhost dracut[1435]: memstrack is not available Nov 27 01:43:09 localhost dracut[1435]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Nov 27 01:43:09 localhost dracut[1435]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'rngd' will not be installed, because command 'rngd' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Nov 27 01:43:09 localhost dracut[1435]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Nov 27 01:43:09 localhost dracut[1435]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Nov 27 01:43:09 localhost dracut[1435]: memstrack is not available Nov 27 01:43:09 localhost dracut[1435]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Nov 27 01:43:09 localhost dracut[1435]: *** Including module: systemd *** Nov 27 01:43:09 localhost dracut[1435]: *** Including module: systemd-initrd *** Nov 27 01:43:09 localhost dracut[1435]: *** Including module: i18n *** Nov 27 01:43:09 localhost dracut[1435]: No KEYMAP configured. Nov 27 01:43:10 localhost dracut[1435]: *** Including module: drm *** Nov 27 01:43:10 localhost dracut[1435]: *** Including module: prefixdevname *** Nov 27 01:43:10 localhost dracut[1435]: *** Including module: kernel-modules *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: kernel-modules-extra *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: qemu *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: fstab-sys *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: rootfs-block *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: terminfo *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: udev-rules *** Nov 27 01:43:11 localhost dracut[1435]: Skipping udev rule: 91-permissions.rules Nov 27 01:43:11 localhost dracut[1435]: Skipping udev rule: 80-drivers-modprobe.rules Nov 27 01:43:11 localhost dracut[1435]: *** Including module: virtiofs *** Nov 27 01:43:11 localhost dracut[1435]: *** Including module: dracut-systemd *** Nov 27 01:43:12 localhost dracut[1435]: *** Including module: usrmount *** Nov 27 01:43:12 localhost dracut[1435]: *** Including module: base *** Nov 27 01:43:12 localhost systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. Nov 27 01:43:12 localhost dracut[1435]: *** Including module: fs-lib *** Nov 27 01:43:12 localhost dracut[1435]: *** Including module: kdumpbase *** Nov 27 01:43:12 localhost dracut[1435]: *** Including module: microcode_ctl-fw_dir_override *** Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl module: mangling fw_dir Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-2d-07"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-2d-07" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4e-03"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-4e-03" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-4f-01" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-55-04"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-55-04" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-5e-03"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-5e-03" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8c-01"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-8c-01" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-0xca"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-8e-9e-0x-0xca" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-dell"... Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: configuration "intel-06-8e-9e-0x-dell" is ignored Nov 27 01:43:12 localhost dracut[1435]: microcode_ctl: final fw_dir: "/lib/firmware/updates/5.14.0-284.11.1.el9_2.x86_64 /lib/firmware/updates /lib/firmware/5.14.0-284.11.1.el9_2.x86_64 /lib/firmware" Nov 27 01:43:12 localhost dracut[1435]: *** Including module: shutdown *** Nov 27 01:43:12 localhost dracut[1435]: *** Including module: squash *** Nov 27 01:43:13 localhost dracut[1435]: *** Including modules done *** Nov 27 01:43:13 localhost dracut[1435]: *** Installing kernel module dependencies *** Nov 27 01:43:13 localhost dracut[1435]: *** Installing kernel module dependencies done *** Nov 27 01:43:13 localhost dracut[1435]: *** Resolving executable dependencies *** Nov 27 01:43:14 localhost dracut[1435]: *** Resolving executable dependencies done *** Nov 27 01:43:14 localhost dracut[1435]: *** Hardlinking files *** Nov 27 01:43:14 localhost dracut[1435]: Mode: real Nov 27 01:43:14 localhost dracut[1435]: Files: 1099 Nov 27 01:43:14 localhost dracut[1435]: Linked: 3 files Nov 27 01:43:14 localhost dracut[1435]: Compared: 0 xattrs Nov 27 01:43:14 localhost dracut[1435]: Compared: 373 files Nov 27 01:43:14 localhost dracut[1435]: Saved: 61.04 KiB Nov 27 01:43:14 localhost dracut[1435]: Duration: 0.046713 seconds Nov 27 01:43:14 localhost dracut[1435]: *** Hardlinking files done *** Nov 27 01:43:14 localhost dracut[1435]: Could not find 'strip'. Not stripping the initramfs. Nov 27 01:43:14 localhost dracut[1435]: *** Generating early-microcode cpio image *** Nov 27 01:43:14 localhost dracut[1435]: *** Constructing AuthenticAMD.bin *** Nov 27 01:43:14 localhost dracut[1435]: *** Store current command line parameters *** Nov 27 01:43:14 localhost dracut[1435]: Stored kernel commandline: Nov 27 01:43:14 localhost dracut[1435]: No dracut internal kernel commandline stored in the initramfs Nov 27 01:43:15 localhost dracut[1435]: *** Install squash loader *** Nov 27 01:43:15 localhost dracut[1435]: *** Squashing the files inside the initramfs *** Nov 27 01:43:16 localhost dracut[1435]: *** Squashing the files inside the initramfs done *** Nov 27 01:43:16 localhost dracut[1435]: *** Creating image file '/boot/initramfs-5.14.0-284.11.1.el9_2.x86_64kdump.img' *** Nov 27 01:43:16 localhost dracut[1435]: *** Creating initramfs image file '/boot/initramfs-5.14.0-284.11.1.el9_2.x86_64kdump.img' done *** Nov 27 01:43:17 localhost kdumpctl[1135]: kdump: kexec: loaded kdump kernel Nov 27 01:43:17 localhost kdumpctl[1135]: kdump: Starting kdump: [OK] Nov 27 01:43:17 localhost systemd[1]: Finished Crash recovery kernel arming. Nov 27 01:43:17 localhost systemd[1]: Startup finished in 1.273s (kernel) + 2.686s (initrd) + 23.543s (userspace) = 27.504s. Nov 27 01:43:31 localhost systemd[1]: systemd-hostnamed.service: Deactivated successfully. Nov 27 01:43:58 localhost sshd[4175]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:43:58 localhost systemd[1]: Created slice User Slice of UID 1000. Nov 27 01:43:58 localhost systemd[1]: Starting User Runtime Directory /run/user/1000... Nov 27 01:43:58 localhost systemd-logind[761]: New session 1 of user zuul. Nov 27 01:43:58 localhost systemd[1]: Finished User Runtime Directory /run/user/1000. Nov 27 01:43:58 localhost systemd[1]: Starting User Manager for UID 1000... Nov 27 01:43:59 localhost systemd[4179]: Queued start job for default target Main User Target. Nov 27 01:43:59 localhost systemd[4179]: Created slice User Application Slice. Nov 27 01:43:59 localhost systemd[4179]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 01:43:59 localhost systemd[4179]: Started Daily Cleanup of User's Temporary Directories. Nov 27 01:43:59 localhost systemd[4179]: Reached target Paths. Nov 27 01:43:59 localhost systemd[4179]: Reached target Timers. Nov 27 01:43:59 localhost systemd[4179]: Starting D-Bus User Message Bus Socket... Nov 27 01:43:59 localhost systemd[4179]: Starting Create User's Volatile Files and Directories... Nov 27 01:43:59 localhost systemd[4179]: Finished Create User's Volatile Files and Directories. Nov 27 01:43:59 localhost systemd[4179]: Listening on D-Bus User Message Bus Socket. Nov 27 01:43:59 localhost systemd[4179]: Reached target Sockets. Nov 27 01:43:59 localhost systemd[4179]: Reached target Basic System. Nov 27 01:43:59 localhost systemd[4179]: Reached target Main User Target. Nov 27 01:43:59 localhost systemd[4179]: Startup finished in 121ms. Nov 27 01:43:59 localhost systemd[1]: Started User Manager for UID 1000. Nov 27 01:43:59 localhost systemd[1]: Started Session 1 of User zuul. Nov 27 01:43:59 localhost python3[4231]: ansible-setup Invoked with gather_subset=['!all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 01:44:11 localhost python3[4249]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 01:44:17 localhost python3[4303]: ansible-setup Invoked with gather_subset=['network'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 01:44:18 localhost python3[4333]: ansible-zuul_console Invoked with path=/tmp/console-{log_uuid}.log port=19885 state=present Nov 27 01:44:21 localhost python3[4349]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:22 localhost python3[4363]: ansible-file Invoked with state=directory path=/home/zuul/.ssh mode=448 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:23 localhost python3[4422]: ansible-ansible.legacy.stat Invoked with path=/home/zuul/.ssh/id_rsa follow=False get_checksum=False checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:44:24 localhost python3[4463]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764225863.5206513-395-48453860487461/source dest=/home/zuul/.ssh/id_rsa mode=384 force=False _original_basename=77016a48c6d24f74861211155030ddad_id_rsa follow=False checksum=ce3ce8a8369bad4c8644d6b47839406e882610f3 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:25 localhost python3[4536]: ansible-ansible.legacy.stat Invoked with path=/home/zuul/.ssh/id_rsa.pub follow=False get_checksum=False checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:44:25 localhost python3[4577]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764225865.1699564-493-236657348220965/source dest=/home/zuul/.ssh/id_rsa.pub mode=420 force=False _original_basename=77016a48c6d24f74861211155030ddad_id_rsa.pub follow=False checksum=a97ae29aa4769480ecb6dcb26255a14feb9b65fa backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:27 localhost python3[4605]: ansible-ping Invoked with data=pong Nov 27 01:44:29 localhost python3[4619]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 01:44:32 localhost python3[4672]: ansible-zuul_debug_info Invoked with ipv4_route_required=False ipv6_route_required=False image_manifest_files=['/etc/dib-builddate.txt', '/etc/image-hostname.txt'] image_manifest=None traceroute_host=None Nov 27 01:44:34 localhost python3[4694]: ansible-file Invoked with path=/home/zuul/zuul-output/logs state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:35 localhost python3[4708]: ansible-file Invoked with path=/home/zuul/zuul-output/artifacts state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:35 localhost python3[4722]: ansible-file Invoked with path=/home/zuul/zuul-output/docs state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:36 localhost python3[4736]: ansible-file Invoked with path=/home/zuul/zuul-output/logs state=directory mode=493 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:36 localhost python3[4750]: ansible-file Invoked with path=/home/zuul/zuul-output/artifacts state=directory mode=493 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:36 localhost python3[4764]: ansible-file Invoked with path=/home/zuul/zuul-output/docs state=directory mode=493 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:39 localhost python3[4780]: ansible-file Invoked with path=/etc/ci state=directory owner=root group=root mode=493 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:40 localhost python3[4828]: ansible-ansible.legacy.stat Invoked with path=/etc/ci/mirror_info.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:44:41 localhost python3[4871]: ansible-ansible.legacy.copy Invoked with dest=/etc/ci/mirror_info.sh owner=root group=root mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764225880.7053845-104-77335268096060/source follow=False _original_basename=mirror_info.sh.j2 checksum=92d92a03afdddee82732741071f662c729080c35 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:48 localhost python3[4899]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEA4Z/c9osaGGtU6X8fgELwfj/yayRurfcKA0HMFfdpPxev2dbwljysMuzoVp4OZmW1gvGtyYPSNRvnzgsaabPNKNo2ym5NToCP6UM+KSe93aln4BcM/24mXChYAbXJQ5Bqq/pIzsGs/pKetQN+vwvMxLOwTvpcsCJBXaa981RKML6xj9l/UZ7IIq1HSEKMvPLxZMWdu0Ut8DkCd5F4nOw9Wgml2uYpDCj5LLCrQQ9ChdOMz8hz6SighhNlRpPkvPaet3OXxr/ytFMu7j7vv06CaEnuMMiY2aTWN1Imin9eHAylIqFHta/3gFfQSWt9jXM7owkBLKL7ATzhaAn+fjNupw== arxcruz@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:49 localhost python3[4913]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDS4Fn6k4deCnIlOtLWqZJyksbepjQt04j8Ed8CGx9EKkj0fKiAxiI4TadXQYPuNHMixZy4Nevjb6aDhL5Z906TfvNHKUrjrG7G26a0k8vdc61NEQ7FmcGMWRLwwc6ReDO7lFpzYKBMk4YqfWgBuGU/K6WLKiVW2cVvwIuGIaYrE1OiiX0iVUUk7KApXlDJMXn7qjSYynfO4mF629NIp8FJal38+Kv+HA+0QkE5Y2xXnzD4Lar5+keymiCHRntPppXHeLIRzbt0gxC7v3L72hpQ3BTBEzwHpeS8KY+SX1y5lRMN45thCHfJqGmARJREDjBvWG8JXOPmVIKQtZmVcD5b mandreou@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:49 localhost python3[4927]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC9MiLfy30deHA7xPOAlew5qUq3UP2gmRMYJi8PtkjFB20/DKeWwWNnkZPqP9AayruRoo51SIiVg870gbZE2jYl+Ncx/FYDe56JeC3ySZsXoAVkC9bP7gkOGqOmJjirvAgPMI7bogVz8i+66Q4Ar7OKTp3762G4IuWPPEg4ce4Y7lx9qWocZapHYq4cYKMxrOZ7SEbFSATBbe2bPZAPKTw8do/Eny+Hq/LkHFhIeyra6cqTFQYShr+zPln0Cr+ro/pDX3bB+1ubFgTpjpkkkQsLhDfR6cCdCWM2lgnS3BTtYj5Ct9/JRPR5YOphqZz+uB+OEu2IL68hmU9vNTth1KeX rlandy@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:49 localhost python3[4941]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFCbgz8gdERiJlk2IKOtkjQxEXejrio6ZYMJAVJYpOIp raukadah@gmail.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:49 localhost python3[4955]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBqb3Q/9uDf4LmihQ7xeJ9gA/STIQUFPSfyyV0m8AoQi bshewale@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:50 localhost python3[4969]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC0I8QqQx0Az2ysJt2JuffucLijhBqnsXKEIx5GyHwxVULROa8VtNFXUDH6ZKZavhiMcmfHB2+TBTda+lDP4FldYj06dGmzCY+IYGa+uDRdxHNGYjvCfLFcmLlzRK6fNbTcui+KlUFUdKe0fb9CRoGKyhlJD5GRkM1Dv+Yb6Bj+RNnmm1fVGYxzmrD2utvffYEb0SZGWxq2R9gefx1q/3wCGjeqvufEV+AskPhVGc5T7t9eyZ4qmslkLh1/nMuaIBFcr9AUACRajsvk6mXrAN1g3HlBf2gQlhi1UEyfbqIQvzzFtsbLDlSum/KmKjy818GzvWjERfQ0VkGzCd9bSLVL dviroel@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:50 localhost python3[4983]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDLOQd4ZLtkZXQGY6UwAr/06ppWQK4fDO3HaqxPk98csyOCBXsliSKK39Bso828+5srIXiW7aI6aC9P5mwi4mUZlGPfJlQbfrcGvY+b/SocuvaGK+1RrHLoJCT52LBhwgrzlXio2jeksZeein8iaTrhsPrOAs7KggIL/rB9hEiB3NaOPWhhoCP4vlW6MEMExGcqB/1FVxXFBPnLkEyW0Lk7ycVflZl2ocRxbfjZi0+tI1Wlinp8PvSQSc/WVrAcDgKjc/mB4ODPOyYy3G8FHgfMsrXSDEyjBKgLKMsdCrAUcqJQWjkqXleXSYOV4q3pzL+9umK+q/e3P/bIoSFQzmJKTU1eDfuvPXmow9F5H54fii/Da7ezlMJ+wPGHJrRAkmzvMbALy7xwswLhZMkOGNtRcPqaKYRmIBKpw3o6bCTtcNUHOtOQnzwY8JzrM2eBWJBXAANYw+9/ho80JIiwhg29CFNpVBuHbql2YxJQNrnl90guN65rYNpDxdIluweyUf8= anbanerj@kaermorhen manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:50 localhost python3[4997]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC3VwV8Im9kRm49lt3tM36hj4Zv27FxGo4C1Q/0jqhzFmHY7RHbmeRr8ObhwWoHjXSozKWg8FL5ER0z3hTwL0W6lez3sL7hUaCmSuZmG5Hnl3x4vTSxDI9JZ/Y65rtYiiWQo2fC5xJhU/4+0e5e/pseCm8cKRSu+SaxhO+sd6FDojA2x1BzOzKiQRDy/1zWGp/cZkxcEuB1wHI5LMzN03c67vmbu+fhZRAUO4dQkvcnj2LrhQtpa+ytvnSjr8icMDosf1OsbSffwZFyHB/hfWGAfe0eIeSA2XPraxiPknXxiPKx2MJsaUTYbsZcm3EjFdHBBMumw5rBI74zLrMRvCO9GwBEmGT4rFng1nP+yw5DB8sn2zqpOsPg1LYRwCPOUveC13P6pgsZZPh812e8v5EKnETct+5XI3dVpdw6CnNiLwAyVAF15DJvBGT/u1k0Myg/bQn+Gv9k2MSj6LvQmf6WbZu2Wgjm30z3FyCneBqTL7mLF19YXzeC0ufHz5pnO1E= dasm@fedora manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:50 localhost python3[5011]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHUnwjB20UKmsSed9X73eGNV5AOEFccQ3NYrRW776pEk cjeanner manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:51 localhost python3[5025]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDercCMGn8rW1C4P67tHgtflPdTeXlpyUJYH+6XDd2lR jgilaber@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:51 localhost python3[5039]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIAMI6kkg9Wg0sG7jIJmyZemEBwUn1yzNpQQd3gnulOmZ adrianfuscoarnejo@gmail.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:51 localhost python3[5053]: ansible-authorized_key Invoked with user=zuul state=present key=ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPijwpQu/3jhhhBZInXNOLEH57DrknPc3PLbsRvYyJIFzwYjX+WD4a7+nGnMYS42MuZk6TJcVqgnqofVx4isoD4= ramishra@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:51 localhost python3[5067]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGpU/BepK3qX0NRf5Np+dOBDqzQEefhNrw2DCZaH3uWW rebtoor@monolith manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:52 localhost python3[5081]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDK0iKdi8jQTpQrDdLVH/AAgLVYyTXF7AQ1gjc/5uT3t ykarel@yatinkarel manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:52 localhost python3[5095]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIF/V/cLotA6LZeO32VL45Hd78skuA2lJA425Sm2LlQeZ fmount@horcrux manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:52 localhost python3[5109]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDa7QCjuDMVmRPo1rREbGwzYeBCYVN+Ou/3WKXZEC6Sr manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:52 localhost python3[5123]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQCfNtF7NvKl915TGsGGoseUb06Hj8L/S4toWf0hExeY+F00woL6NvBlJD0nDct+P5a22I4EhvoQCRQ8reaPCm1lybR3uiRIJsj+8zkVvLwby9LXzfZorlNG9ofjd00FEmB09uW/YvTl6Q9XwwwX6tInzIOv3TMqTHHGOL74ibbj8J/FJR0cFEyj0z4WQRvtkh32xAHl83gbuINryMt0sqRI+clj2381NKL55DRLQrVw0gsfqqxiHAnXg21qWmc4J+b9e9kiuAFQjcjwTVkwJCcg3xbPwC/qokYRby/Y5S40UUd7/jEARGXT7RZgpzTuDd1oZiCVrnrqJNPaMNdVv5MLeFdf1B7iIe5aa/fGouX7AO4SdKhZUdnJmCFAGvjC6S3JMZ2wAcUl+OHnssfmdj7XL50cLo27vjuzMtLAgSqi6N99m92WCF2s8J9aVzszX7Xz9OKZCeGsiVJp3/NdABKzSEAyM9xBD/5Vho894Sav+otpySHe3p6RUTgbB5Zu8VyZRZ/UtB3ueXxyo764yrc6qWIDqrehm84Xm9g+/jpIBzGPl07NUNJpdt/6Sgf9RIKXw/7XypO5yZfUcuFNGTxLfqjTNrtgLZNcjfav6sSdVXVcMPL//XNuRdKmVFaO76eV/oGMQGr1fGcCD+N+CpI7+Q+fCNB6VFWG4nZFuI/Iuw== averdagu@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:53 localhost python3[5137]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDq8l27xI+QlQVdS4djp9ogSoyrNE2+Ox6vKPdhSNL1J3PE5w+WCSvMz9A5gnNuH810zwbekEApbxTze/gLQJwBHA52CChfURpXrFaxY7ePXRElwKAL3mJfzBWY/c5jnNL9TCVmFJTGZkFZP3Nh+BMgZvL6xBkt3WKm6Uq18qzd9XeKcZusrA+O+uLv1fVeQnadY9RIqOCyeFYCzLWrUfTyE8x/XG0hAWIM7qpnF2cALQS2h9n4hW5ybiUN790H08wf9hFwEf5nxY9Z9dVkPFQiTSGKNBzmnCXU9skxS/xhpFjJ5duGSZdtAHe9O+nGZm9c67hxgtf8e5PDuqAdXEv2cf6e3VBAt+Bz8EKI3yosTj0oZHfwr42Yzb1l/SKy14Rggsrc9KAQlrGXan6+u2jcQqqx7l+SWmnpFiWTV9u5cWj2IgOhApOitmRBPYqk9rE2usfO0hLn/Pj/R/Nau4803e1/EikdLE7Ps95s9mX5jRDjAoUa2JwFF5RsVFyL910= ashigupt@ashigupt.remote.csb manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:53 localhost python3[5151]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOKLl0NYKwoZ/JY5KeZU8VwRAggeOxqQJeoqp3dsAaY9 manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:53 localhost python3[5165]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIASASQOH2BcOyLKuuDOdWZlPi2orcjcA8q4400T73DLH evallesp@fedora manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:53 localhost python3[5179]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILeBWlamUph+jRKV2qrx1PGU7vWuGIt5+z9k96I8WehW amsinha@amsinha-mac manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:54 localhost python3[5193]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIANvVgvJBlK3gb1yz5uef/JqIGq4HLEmY2dYA8e37swb morenod@redhat-laptop manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:54 localhost python3[5207]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQDZdI7t1cxYx65heVI24HTV4F7oQLW1zyfxHreL2TIJKxjyrUUKIFEUmTutcBlJRLNT2Eoix6x1sOw9YrchloCLcn//SGfTElr9mSc5jbjb7QXEU+zJMhtxyEJ1Po3CUGnj7ckiIXw7wcawZtrEOAQ9pH3ExYCJcEMiyNjRQZCxT3tPK+S4B95EWh5Fsrz9CkwpjNRPPH7LigCeQTM3Wc7r97utAslBUUvYceDSLA7rMgkitJE38b7rZBeYzsGQ8YYUBjTCtehqQXxCRjizbHWaaZkBU+N3zkKB6n/iCNGIO690NK7A/qb6msTijiz1PeuM8ThOsi9qXnbX5v0PoTpcFSojV7NHAQ71f0XXuS43FhZctT+Dcx44dT8Fb5vJu2cJGrk+qF8ZgJYNpRS7gPg0EG2EqjK7JMf9ULdjSu0r+KlqIAyLvtzT4eOnQipoKlb/WG5D/0ohKv7OMQ352ggfkBFIQsRXyyTCT98Ft9juqPuahi3CAQmP4H9dyE+7+Kz437PEtsxLmfm6naNmWi7Ee1DqWPwS8rEajsm4sNM4wW9gdBboJQtc0uZw0DfLj1I9r3Mc8Ol0jYtz0yNQDSzVLrGCaJlC311trU70tZ+ZkAVV6Mn8lOhSbj1cK0lvSr6ZK4dgqGl3I1eTZJJhbLNdg7UOVaiRx9543+C/p/As7w== brjackma@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:54 localhost python3[5221]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKwedoZ0TWPJX/z/4TAbO/kKcDZOQVgRH0hAqrL5UCI1 vcastell@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:54 localhost python3[5235]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEmv8sE8GCk6ZTPIqF0FQrttBdL3mq7rCm/IJy0xDFh7 michburk@redhat.com manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:55 localhost python3[5249]: ansible-authorized_key Invoked with user=zuul state=present key=ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICy6GpGEtwevXEEn4mmLR5lmSLe23dGgAvzkB9DMNbkf rsafrono@rsafrono manage_dir=True exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 01:44:56 localhost python3[5265]: ansible-community.general.timezone Invoked with name=UTC hwclock=None Nov 27 01:44:56 localhost systemd[1]: Starting Time & Date Service... Nov 27 01:44:57 localhost systemd[1]: Started Time & Date Service. Nov 27 01:44:57 localhost systemd-timedated[5267]: Changed time zone to 'UTC' (UTC). Nov 27 01:44:58 localhost python3[5287]: ansible-file Invoked with path=/etc/nodepool state=directory mode=511 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:44:59 localhost python3[5333]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/sub_nodes follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:45:00 localhost python3[5374]: ansible-ansible.legacy.copy Invoked with dest=/etc/nodepool/sub_nodes src=/home/zuul/.ansible/tmp/ansible-tmp-1764225899.6514099-501-61424667616432/source _original_basename=tmpzsgh40zv follow=False checksum=da39a3ee5e6b4b0d3255bfef95601890afd80709 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:45:01 localhost python3[5434]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/sub_nodes_private follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:45:01 localhost python3[5475]: ansible-ansible.legacy.copy Invoked with dest=/etc/nodepool/sub_nodes_private src=/home/zuul/.ansible/tmp/ansible-tmp-1764225901.077536-584-48552703685591/source _original_basename=tmp6_a7_53l follow=False checksum=da39a3ee5e6b4b0d3255bfef95601890afd80709 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:45:03 localhost python3[5537]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/node_private follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:45:03 localhost python3[5580]: ansible-ansible.legacy.copy Invoked with dest=/etc/nodepool/node_private src=/home/zuul/.ansible/tmp/ansible-tmp-1764225903.224636-730-64740296767671/source _original_basename=tmprjy0xafk follow=False checksum=9002ae785196258bce68f82c9276ee1756ef1744 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:45:04 localhost python3[5608]: ansible-ansible.legacy.command Invoked with _raw_params=cp .ssh/id_rsa /etc/nodepool/id_rsa zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:45:05 localhost python3[5624]: ansible-ansible.legacy.command Invoked with _raw_params=cp .ssh/id_rsa.pub /etc/nodepool/id_rsa.pub zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:45:06 localhost python3[5674]: ansible-ansible.legacy.stat Invoked with path=/etc/sudoers.d/zuul-sudo-grep follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:45:06 localhost python3[5717]: ansible-ansible.legacy.copy Invoked with dest=/etc/sudoers.d/zuul-sudo-grep mode=288 src=/home/zuul/.ansible/tmp/ansible-tmp-1764225906.1871138-857-278385974082808/source _original_basename=tmptz3sqite follow=False checksum=bdca1a77493d00fb51567671791f4aa30f66c2f0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:45:18 localhost python3[5748]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/visudo -c zuul_log_id=fa163ec2-ffbe-9d9e-4507-000000000023-1-overcloudnovacompute2 zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:45:27 localhost systemd[1]: systemd-timedated.service: Deactivated successfully. Nov 27 01:45:29 localhost python3[5768]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=env#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-9d9e-4507-000000000024-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Nov 27 01:45:30 localhost python3[5786]: ansible-file Invoked with path=/home/zuul/workspace state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:45:50 localhost python3[5803]: ansible-ansible.builtin.file Invoked with path=/etc/ci/env state=directory mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:46:34 localhost systemd[4179]: Starting Mark boot as successful... Nov 27 01:46:34 localhost systemd[4179]: Finished Mark boot as successful. Nov 27 01:46:50 localhost systemd-logind[761]: Session 1 logged out. Waiting for processes to exit. Nov 27 01:46:58 localhost systemd[1]: Unmounting EFI System Partition Automount... Nov 27 01:46:58 localhost systemd[1]: efi.mount: Deactivated successfully. Nov 27 01:46:58 localhost systemd[1]: Unmounted EFI System Partition Automount. Nov 27 01:49:34 localhost systemd[4179]: Created slice User Background Tasks Slice. Nov 27 01:49:34 localhost systemd[4179]: Starting Cleanup of User's Temporary Files and Directories... Nov 27 01:49:34 localhost systemd[4179]: Finished Cleanup of User's Temporary Files and Directories. Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: [1af4:1000] type 00 class 0x020000 Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: reg 0x10: [io 0x0000-0x003f] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: reg 0x14: [mem 0x00000000-0x00000fff] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: reg 0x20: [mem 0x00000000-0x00003fff 64bit pref] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: reg 0x30: [mem 0x00000000-0x0007ffff pref] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: BAR 6: assigned [mem 0xc0000000-0xc007ffff pref] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: BAR 4: assigned [mem 0x440000000-0x440003fff 64bit pref] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: BAR 1: assigned [mem 0xc0080000-0xc0080fff] Nov 27 01:49:50 localhost kernel: pci 0000:00:07.0: BAR 0: assigned [io 0x1000-0x103f] Nov 27 01:49:50 localhost kernel: virtio-pci 0000:00:07.0: enabling device (0000 -> 0003) Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1655] manager: (eth1): new Ethernet device (/org/freedesktop/NetworkManager/Devices/3) Nov 27 01:49:50 localhost systemd-udevd[5811]: Network interface NamePolicy= disabled on kernel command line. Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1773] device (eth1): state change: unmanaged -> unavailable (reason 'managed', sys-iface-state: 'external') Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1795] settings: (eth1): created default wired connection 'Wired connection 1' Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1798] device (eth1): carrier: link connected Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1800] device (eth1): state change: unavailable -> disconnected (reason 'carrier-changed', sys-iface-state: 'managed') Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1805] policy: auto-activating connection 'Wired connection 1' (636f7d9a-c0f3-3132-acf9-5630c320f73e) Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1809] device (eth1): Activation: starting connection 'Wired connection 1' (636f7d9a-c0f3-3132-acf9-5630c320f73e) Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1810] device (eth1): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'managed') Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1814] device (eth1): state change: prepare -> config (reason 'none', sys-iface-state: 'managed') Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1819] device (eth1): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') Nov 27 01:49:50 localhost NetworkManager[790]: [1764226190.1823] dhcp4 (eth1): activation: beginning transaction (timeout in 45 seconds) Nov 27 01:49:51 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth1: link becomes ready Nov 27 01:49:51 localhost sshd[5813]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:49:51 localhost systemd-logind[761]: New session 3 of user zuul. Nov 27 01:49:51 localhost systemd[1]: Started Session 3 of User zuul. Nov 27 01:49:51 localhost python3[5830]: ansible-ansible.legacy.command Invoked with _raw_params=ip -j link zuul_log_id=fa163ec2-ffbe-5241-932c-000000000475-0-controller zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:50:04 localhost python3[5880]: ansible-ansible.legacy.stat Invoked with path=/etc/NetworkManager/system-connections/ci-private-network.nmconnection follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:50:05 localhost python3[5923]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764226204.5857074-537-30036674308233/source dest=/etc/NetworkManager/system-connections/ci-private-network.nmconnection mode=0600 owner=root group=root follow=False _original_basename=bootstrap-ci-network-nm-connection.nmconnection.j2 checksum=b4bc9a90ead39e1b236999bb7ab6259ed6487450 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:50:05 localhost python3[5953]: ansible-ansible.builtin.systemd Invoked with name=NetworkManager state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 01:50:05 localhost systemd[1]: NetworkManager-wait-online.service: Deactivated successfully. Nov 27 01:50:05 localhost systemd[1]: Stopped Network Manager Wait Online. Nov 27 01:50:05 localhost systemd[1]: Stopping Network Manager Wait Online... Nov 27 01:50:05 localhost systemd[1]: Stopping Network Manager... Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7591] caught SIGTERM, shutting down normally. Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7685] dhcp4 (eth0): canceled DHCP transaction Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7686] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7687] dhcp4 (eth0): state changed no lease Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7695] manager: NetworkManager state is now CONNECTING Nov 27 01:50:05 localhost systemd[1]: Starting Network Manager Script Dispatcher Service... Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7835] dhcp4 (eth1): canceled DHCP transaction Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7836] dhcp4 (eth1): state changed no lease Nov 27 01:50:05 localhost systemd[1]: Started Network Manager Script Dispatcher Service. Nov 27 01:50:05 localhost NetworkManager[790]: [1764226205.7942] exiting (success) Nov 27 01:50:05 localhost systemd[1]: NetworkManager.service: Deactivated successfully. Nov 27 01:50:05 localhost systemd[1]: Stopped Network Manager. Nov 27 01:50:05 localhost systemd[1]: NetworkManager.service: Consumed 2.319s CPU time. Nov 27 01:50:05 localhost systemd[1]: Starting Network Manager... Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.8407] NetworkManager (version 1.42.2-1.el9) is starting... (after a restart, boot:c97b6fe2-88a9-4c40-99a9-d67ccebb1335) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.8409] Read config: /etc/NetworkManager/NetworkManager.conf (run: 15-carrier-timeout.conf) Nov 27 01:50:05 localhost systemd[1]: Started Network Manager. Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.8446] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Nov 27 01:50:05 localhost systemd[1]: Starting Network Manager Wait Online... Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.8497] manager[0x55b36e8ba090]: monitoring kernel firmware directory '/lib/firmware'. Nov 27 01:50:05 localhost systemd[1]: Starting Hostname Service... Nov 27 01:50:05 localhost systemd[1]: Started Hostname Service. Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9263] hostname: hostname: using hostnamed Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9263] hostname: static hostname changed from (none) to "np0005537446.novalocal" Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9269] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9276] manager[0x55b36e8ba090]: rfkill: Wi-Fi hardware radio set enabled Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9276] manager[0x55b36e8ba090]: rfkill: WWAN hardware radio set enabled Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9313] Loaded device plugin: NMTeamFactory (/usr/lib64/NetworkManager/1.42.2-1.el9/libnm-device-plugin-team.so) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9313] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9314] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9315] manager: Networking is enabled by state file Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9325] settings: Loaded settings plugin: ifcfg-rh ("/usr/lib64/NetworkManager/1.42.2-1.el9/libnm-settings-plugin-ifcfg-rh.so") Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9325] settings: Loaded settings plugin: keyfile (internal) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9367] dhcp: init: Using DHCP client 'internal' Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9371] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9378] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9386] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9398] device (lo): Activation: starting connection 'lo' (10cb86f0-06ff-4d9e-a0e4-caa044d1285f) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9406] device (eth0): carrier: link connected Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9411] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9418] manager: (eth0): assume: will attempt to assume matching connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) (indicated) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9419] device (eth0): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9426] device (eth0): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9436] device (eth0): Activation: starting connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9443] device (eth1): carrier: link connected Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9449] manager: (eth1): new Ethernet device (/org/freedesktop/NetworkManager/Devices/3) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9456] manager: (eth1): assume: will attempt to assume matching connection 'Wired connection 1' (636f7d9a-c0f3-3132-acf9-5630c320f73e) (indicated) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9457] device (eth1): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9464] device (eth1): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9473] device (eth1): Activation: starting connection 'Wired connection 1' (636f7d9a-c0f3-3132-acf9-5630c320f73e) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9500] device (lo): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9508] device (lo): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9510] device (lo): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9513] device (eth0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9517] device (eth0): state change: prepare -> config (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9520] device (eth1): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9523] device (eth1): state change: prepare -> config (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9551] device (lo): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9558] device (eth0): state change: config -> ip-config (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9562] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9572] device (eth1): state change: config -> ip-config (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9577] dhcp4 (eth1): activation: beginning transaction (timeout in 45 seconds) Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9593] device (lo): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9599] device (lo): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9607] device (lo): Activation: successful, device activated. Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9616] dhcp4 (eth0): state changed new lease, address=38.102.83.107 Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9622] policy: set 'System eth0' (eth0) as default for IPv4 routing and DNS Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9748] device (eth0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9786] device (eth0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9788] device (eth0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9793] manager: NetworkManager state is now CONNECTED_SITE Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9797] device (eth0): Activation: successful, device activated. Nov 27 01:50:05 localhost NetworkManager[5971]: [1764226205.9803] manager: NetworkManager state is now CONNECTED_GLOBAL Nov 27 01:50:06 localhost python3[6023]: ansible-ansible.legacy.command Invoked with _raw_params=ip route zuul_log_id=fa163ec2-ffbe-5241-932c-000000000136-0-controller zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:50:16 localhost systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. Nov 27 01:50:35 localhost systemd[1]: systemd-hostnamed.service: Deactivated successfully. Nov 27 01:50:50 localhost NetworkManager[5971]: [1764226250.8127] device (eth1): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:50 localhost systemd[1]: Starting Network Manager Script Dispatcher Service... Nov 27 01:50:50 localhost systemd[1]: Started Network Manager Script Dispatcher Service. Nov 27 01:50:50 localhost NetworkManager[5971]: [1764226250.8321] device (eth1): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:50 localhost NetworkManager[5971]: [1764226250.8325] device (eth1): state change: secondaries -> activated (reason 'none', sys-iface-state: 'assume') Nov 27 01:50:50 localhost NetworkManager[5971]: [1764226250.8331] device (eth1): Activation: successful, device activated. Nov 27 01:50:50 localhost NetworkManager[5971]: [1764226250.8338] manager: startup complete Nov 27 01:50:50 localhost systemd[1]: Finished Network Manager Wait Online. Nov 27 01:51:00 localhost systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. Nov 27 01:51:06 localhost systemd[1]: session-3.scope: Deactivated successfully. Nov 27 01:51:06 localhost systemd[1]: session-3.scope: Consumed 1.388s CPU time. Nov 27 01:51:06 localhost systemd-logind[761]: Session 3 logged out. Waiting for processes to exit. Nov 27 01:51:06 localhost systemd-logind[761]: Removed session 3. Nov 27 01:51:26 localhost sshd[6052]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:51:26 localhost systemd-logind[761]: New session 4 of user zuul. Nov 27 01:51:26 localhost systemd[1]: Started Session 4 of User zuul. Nov 27 01:51:27 localhost python3[6103]: ansible-ansible.legacy.stat Invoked with path=/etc/ci/env/networking-info.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:51:27 localhost python3[6146]: ansible-ansible.legacy.copy Invoked with dest=/etc/ci/env/networking-info.yml owner=root group=root mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764226286.9255815-628-84510252682398/source _original_basename=tmpwk_z6zsg follow=False checksum=80907ec8159b460a1a8a6328349d5fe7c8bfc862 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:51:29 localhost systemd[1]: session-4.scope: Deactivated successfully. Nov 27 01:51:29 localhost systemd-logind[761]: Session 4 logged out. Waiting for processes to exit. Nov 27 01:51:29 localhost systemd-logind[761]: Removed session 4. Nov 27 01:57:34 localhost systemd[1]: Starting dnf makecache... Nov 27 01:57:35 localhost dnf[6162]: Failed determining last makecache time. Nov 27 01:57:35 localhost dnf[6162]: There are no enabled repositories in "/etc/yum.repos.d", "/etc/yum/repos.d", "/etc/distro.repos.d". Nov 27 01:57:35 localhost systemd[1]: dnf-makecache.service: Deactivated successfully. Nov 27 01:57:35 localhost systemd[1]: Finished dnf makecache. Nov 27 01:58:34 localhost systemd[1]: Starting Cleanup of Temporary Directories... Nov 27 01:58:34 localhost systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully. Nov 27 01:58:34 localhost systemd[1]: Finished Cleanup of Temporary Directories. Nov 27 01:58:34 localhost systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dclean.service.mount: Deactivated successfully. Nov 27 01:59:24 localhost sshd[6168]: main: sshd: ssh-rsa algorithm is disabled Nov 27 01:59:24 localhost systemd-logind[761]: New session 5 of user zuul. Nov 27 01:59:24 localhost systemd[1]: Started Session 5 of User zuul. Nov 27 01:59:25 localhost python3[6187]: ansible-ansible.legacy.command Invoked with _raw_params=lsblk -nd -o MAJ:MIN /dev/vda#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-3e40-824e-000000001d12-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:26 localhost python3[6206]: ansible-ansible.builtin.file Invoked with path=/sys/fs/cgroup/init.scope state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:26 localhost python3[6222]: ansible-ansible.builtin.file Invoked with path=/sys/fs/cgroup/machine.slice state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:26 localhost python3[6238]: ansible-ansible.builtin.file Invoked with path=/sys/fs/cgroup/system.slice state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:27 localhost python3[6254]: ansible-ansible.builtin.file Invoked with path=/sys/fs/cgroup/user.slice state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:27 localhost python3[6270]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system.conf.d state=directory mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:29 localhost python3[6318]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system.conf.d/override.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 01:59:29 localhost python3[6361]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system.conf.d/override.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764226768.9491944-655-242341143249640/source _original_basename=tmpge0d4cm5 follow=False checksum=a05098bd3d2321238ea1169d0e6f135b35b392d4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 01:59:31 localhost python3[6391]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 01:59:31 localhost systemd[1]: Reloading. Nov 27 01:59:31 localhost systemd-rc-local-generator[6409]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 01:59:31 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 01:59:32 localhost python3[6438]: ansible-ansible.builtin.wait_for Invoked with path=/sys/fs/cgroup/system.slice/io.max state=present timeout=30 host=127.0.0.1 connect_timeout=5 delay=0 active_connection_states=['ESTABLISHED', 'FIN_WAIT1', 'FIN_WAIT2', 'SYN_RECV', 'SYN_SENT', 'TIME_WAIT'] sleep=1 port=None search_regex=None exclude_hosts=None msg=None Nov 27 01:59:33 localhost python3[6455]: ansible-ansible.legacy.command Invoked with _raw_params=echo "252:0 riops=18000 wiops=18000 rbps=262144000 wbps=262144000" > /sys/fs/cgroup/init.scope/io.max#012 _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:34 localhost python3[6473]: ansible-ansible.legacy.command Invoked with _raw_params=echo "252:0 riops=18000 wiops=18000 rbps=262144000 wbps=262144000" > /sys/fs/cgroup/machine.slice/io.max#012 _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:34 localhost python3[6491]: ansible-ansible.legacy.command Invoked with _raw_params=echo "252:0 riops=18000 wiops=18000 rbps=262144000 wbps=262144000" > /sys/fs/cgroup/system.slice/io.max#012 _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:34 localhost python3[6509]: ansible-ansible.legacy.command Invoked with _raw_params=echo "252:0 riops=18000 wiops=18000 rbps=262144000 wbps=262144000" > /sys/fs/cgroup/user.slice/io.max#012 _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:36 localhost python3[6526]: ansible-ansible.legacy.command Invoked with _raw_params=echo "init"; cat /sys/fs/cgroup/init.scope/io.max; echo "machine"; cat /sys/fs/cgroup/machine.slice/io.max; echo "system"; cat /sys/fs/cgroup/system.slice/io.max; echo "user"; cat /sys/fs/cgroup/user.slice/io.max;#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-3e40-824e-000000001d19-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 01:59:36 localhost python3[6546]: ansible-ansible.builtin.stat Invoked with path=/sys/fs/cgroup/kubepods.slice/io.max follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 01:59:39 localhost systemd[1]: session-5.scope: Deactivated successfully. Nov 27 01:59:39 localhost systemd[1]: session-5.scope: Consumed 3.909s CPU time. Nov 27 01:59:39 localhost systemd-logind[761]: Session 5 logged out. Waiting for processes to exit. Nov 27 01:59:39 localhost systemd-logind[761]: Removed session 5. Nov 27 02:01:01 localhost sshd[6552]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:01:01 localhost systemd-logind[761]: New session 6 of user zuul. Nov 27 02:01:01 localhost systemd[1]: Started Session 6 of User zuul. Nov 27 02:01:02 localhost systemd[1]: Starting RHSM dbus service... Nov 27 02:01:02 localhost systemd[1]: Started RHSM dbus service. Nov 27 02:01:02 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:169] Could not import locale for C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:02 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:139] Could not import locale either for C_C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:02 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:169] Could not import locale for C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:02 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:139] Could not import locale either for C_C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:05 localhost rhsm-service[6591]: INFO [subscription_manager.managerlib:90] Consumer created: np0005537446.novalocal (88fcf67a-1c7b-4599-8a4d-ef1b8885a3ba) Nov 27 02:01:05 localhost subscription-manager[6591]: Registered system with identity: 88fcf67a-1c7b-4599-8a4d-ef1b8885a3ba Nov 27 02:01:05 localhost rhsm-service[6591]: INFO [subscription_manager.entcertlib:131] certs updated: Nov 27 02:01:05 localhost rhsm-service[6591]: Total updates: 1 Nov 27 02:01:05 localhost rhsm-service[6591]: Found (local) serial# [] Nov 27 02:01:05 localhost rhsm-service[6591]: Expected (UEP) serial# [9221875700159912337] Nov 27 02:01:05 localhost rhsm-service[6591]: Added (new) Nov 27 02:01:05 localhost rhsm-service[6591]: [sn:9221875700159912337 ( Content Access,) @ /etc/pki/entitlement/9221875700159912337.pem] Nov 27 02:01:05 localhost rhsm-service[6591]: Deleted (rogue): Nov 27 02:01:05 localhost rhsm-service[6591]: Nov 27 02:01:05 localhost subscription-manager[6591]: Added subscription for 'Content Access' contract 'None' Nov 27 02:01:05 localhost subscription-manager[6591]: Added subscription for product ' Content Access' Nov 27 02:01:07 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:169] Could not import locale for C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:07 localhost rhsm-service[6591]: INFO [subscription_manager.i18n:139] Could not import locale either for C_C: [Errno 2] No translation file found for domain: 'rhsm' Nov 27 02:01:07 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:01:07 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:01:07 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:01:07 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:01:07 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:01:09 localhost python3[6682]: ansible-ansible.legacy.command Invoked with _raw_params=cat /etc/redhat-release zuul_log_id=fa163ec2-ffbe-b427-d4f9-00000000000d-1-overcloudnovacompute2 zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:02:02 localhost python3[6701]: ansible-ansible.legacy.dnf Invoked with name=['podman'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:02:31 localhost setsebool[6776]: The virt_use_nfs policy boolean was changed to 1 by root Nov 27 02:02:31 localhost setsebool[6776]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Nov 27 02:02:39 localhost kernel: SELinux: Converting 410 SID table entries... Nov 27 02:02:39 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:02:39 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:02:39 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:02:39 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:02:39 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:02:39 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:02:39 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:02:52 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=3 res=1 Nov 27 02:02:52 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:02:52 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:02:52 localhost systemd[1]: Reloading. Nov 27 02:02:52 localhost systemd-rc-local-generator[7625]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:02:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:02:52 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:02:54 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:03:01 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:03:01 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:03:01 localhost systemd[1]: man-db-cache-update.service: Consumed 10.030s CPU time. Nov 27 02:03:01 localhost systemd[1]: run-r20924cb4980f4021b56ecddcf4ce4e77.service: Deactivated successfully. Nov 27 02:03:03 localhost podman[18370]: 2025-11-27 07:03:03.015626426 +0000 UTC m=+0.099608633 system refresh Nov 27 02:03:03 localhost systemd[4179]: Starting D-Bus User Message Bus... Nov 27 02:03:03 localhost dbus-broker-launch[18428]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +31: Eavesdropping is deprecated and ignored Nov 27 02:03:03 localhost dbus-broker-launch[18428]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +33: Eavesdropping is deprecated and ignored Nov 27 02:03:03 localhost systemd[4179]: Started D-Bus User Message Bus. Nov 27 02:03:03 localhost journal[18428]: Ready Nov 27 02:03:03 localhost systemd[4179]: selinux: avc: op=load_policy lsm=selinux seqno=3 res=1 Nov 27 02:03:03 localhost systemd[4179]: Created slice Slice /user. Nov 27 02:03:03 localhost systemd[4179]: podman-18411.scope: unit configures an IP firewall, but not running as root. Nov 27 02:03:03 localhost systemd[4179]: (This warning is only shown for the first unit using IP firewalling.) Nov 27 02:03:03 localhost systemd[4179]: Started podman-18411.scope. Nov 27 02:03:03 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:03:04 localhost systemd[4179]: Started podman-pause-8297f363.scope. Nov 27 02:03:04 localhost systemd[1]: session-6.scope: Deactivated successfully. Nov 27 02:03:04 localhost systemd[1]: session-6.scope: Consumed 49.659s CPU time. Nov 27 02:03:04 localhost systemd-logind[761]: Session 6 logged out. Waiting for processes to exit. Nov 27 02:03:04 localhost systemd-logind[761]: Removed session 6. Nov 27 02:03:20 localhost sshd[18434]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:20 localhost sshd[18431]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:20 localhost sshd[18432]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:20 localhost sshd[18433]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:20 localhost sshd[18435]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:25 localhost sshd[18441]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:03:25 localhost systemd-logind[761]: New session 7 of user zuul. Nov 27 02:03:25 localhost systemd[1]: Started Session 7 of User zuul. Nov 27 02:03:25 localhost python3[18458]: ansible-ansible.posix.authorized_key Invoked with user=zuul key=ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEVj7X+JpJUXCYoSUzz1+FID+x4ebTUQFVPIqBYPnXx2LkIL8pW9eSM4W3478cmDLhIF4rwAjxDSyLRRVd4Hw64= zuul@np0005537436.novalocal#012 manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:03:26 localhost python3[18474]: ansible-ansible.posix.authorized_key Invoked with user=root key=ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEVj7X+JpJUXCYoSUzz1+FID+x4ebTUQFVPIqBYPnXx2LkIL8pW9eSM4W3478cmDLhIF4rwAjxDSyLRRVd4Hw64= zuul@np0005537436.novalocal#012 manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:03:27 localhost systemd[1]: session-7.scope: Deactivated successfully. Nov 27 02:03:27 localhost systemd-logind[761]: Session 7 logged out. Waiting for processes to exit. Nov 27 02:03:27 localhost systemd-logind[761]: Removed session 7. Nov 27 02:04:54 localhost sshd[18477]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:04:54 localhost systemd-logind[761]: New session 8 of user zuul. Nov 27 02:04:54 localhost systemd[1]: Started Session 8 of User zuul. Nov 27 02:04:55 localhost python3[18496]: ansible-authorized_key Invoked with user=root manage_dir=True key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:04:56 localhost python3[18512]: ansible-user Invoked with name=root state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.novalocal update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Nov 27 02:04:57 localhost python3[18562]: ansible-ansible.legacy.stat Invoked with path=/root/.ssh/id_rsa follow=False get_checksum=False checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:04:58 localhost python3[18605]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764227097.503051-138-169609717871657/source dest=/root/.ssh/id_rsa mode=384 owner=root force=False _original_basename=77016a48c6d24f74861211155030ddad_id_rsa follow=False checksum=ce3ce8a8369bad4c8644d6b47839406e882610f3 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:04:59 localhost python3[18667]: ansible-ansible.legacy.stat Invoked with path=/root/.ssh/id_rsa.pub follow=False get_checksum=False checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:04:59 localhost python3[18710]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764227099.0992856-231-225432540178931/source dest=/root/.ssh/id_rsa.pub mode=420 owner=root force=False _original_basename=77016a48c6d24f74861211155030ddad_id_rsa.pub follow=False checksum=a97ae29aa4769480ecb6dcb26255a14feb9b65fa backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:05:01 localhost python3[18740]: ansible-ansible.builtin.file Invoked with path=/etc/nodepool state=directory mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:05:02 localhost python3[18786]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/sub_nodes follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:05:03 localhost python3[18802]: ansible-ansible.legacy.file Invoked with dest=/etc/nodepool/sub_nodes _original_basename=tmpom52ovvs recurse=False state=file path=/etc/nodepool/sub_nodes force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:05:04 localhost python3[18862]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/sub_nodes_private follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:05:04 localhost python3[18878]: ansible-ansible.legacy.file Invoked with dest=/etc/nodepool/sub_nodes_private _original_basename=tmpkw6_qnbh recurse=False state=file path=/etc/nodepool/sub_nodes_private force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:05:05 localhost python3[18938]: ansible-ansible.legacy.stat Invoked with path=/etc/nodepool/node_private follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:05:06 localhost python3[18954]: ansible-ansible.legacy.file Invoked with dest=/etc/nodepool/node_private _original_basename=tmptotthff4 recurse=False state=file path=/etc/nodepool/node_private force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:05:06 localhost systemd[1]: session-8.scope: Deactivated successfully. Nov 27 02:05:06 localhost systemd[1]: session-8.scope: Consumed 3.438s CPU time. Nov 27 02:05:06 localhost systemd-logind[761]: Session 8 logged out. Waiting for processes to exit. Nov 27 02:05:06 localhost systemd-logind[761]: Removed session 8. Nov 27 02:06:00 localhost sshd[18969]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:07:18 localhost sshd[18970]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:07:18 localhost systemd-logind[761]: New session 9 of user zuul. Nov 27 02:07:18 localhost systemd[1]: Started Session 9 of User zuul. Nov 27 02:07:18 localhost python3[19016]: ansible-ansible.legacy.command Invoked with _raw_params=hostname _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:12:18 localhost systemd[1]: session-9.scope: Deactivated successfully. Nov 27 02:12:18 localhost systemd-logind[761]: Session 9 logged out. Waiting for processes to exit. Nov 27 02:12:18 localhost systemd-logind[761]: Removed session 9. Nov 27 02:13:34 localhost sshd[19021]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:18:46 localhost sshd[19024]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:18:46 localhost systemd-logind[761]: New session 10 of user zuul. Nov 27 02:18:46 localhost systemd[1]: Started Session 10 of User zuul. Nov 27 02:18:47 localhost python3[19041]: ansible-ansible.legacy.command Invoked with _raw_params=cat /etc/redhat-release zuul_log_id=fa163ec2-ffbe-0cb1-928f-00000000000c-1-overcloudnovacompute2 zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:18:50 localhost python3[19061]: ansible-ansible.legacy.command Invoked with _raw_params=yum clean all zuul_log_id=fa163ec2-ffbe-0cb1-928f-00000000000d-1-overcloudnovacompute2 zuul_ansible_split_streams=False _uses_shell=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:19:20 localhost python3[19081]: ansible-community.general.rhsm_repository Invoked with name=['rhel-9-for-x86_64-baseos-eus-rpms'] state=enabled purge=False Nov 27 02:19:24 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:19:55 localhost python3[19237]: ansible-community.general.rhsm_repository Invoked with name=['rhel-9-for-x86_64-appstream-eus-rpms'] state=enabled purge=False Nov 27 02:19:57 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:16 localhost python3[19378]: ansible-community.general.rhsm_repository Invoked with name=['rhel-9-for-x86_64-highavailability-eus-rpms'] state=enabled purge=False Nov 27 02:20:18 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:23 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:46 localhost python3[19652]: ansible-community.general.rhsm_repository Invoked with name=['fast-datapath-for-rhel-9-x86_64-rpms'] state=enabled purge=False Nov 27 02:20:49 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:49 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:55 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:20:55 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:21:19 localhost python3[20048]: ansible-community.general.rhsm_repository Invoked with name=['openstack-17.1-for-rhel-9-x86_64-rpms'] state=enabled purge=False Nov 27 02:21:22 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:21:22 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:21:27 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:21:27 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:21:38 localhost python3[20445]: ansible-ansible.legacy.command Invoked with _raw_params=yum repolist --enabled#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-000000000013-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:22:04 localhost python3[20464]: ansible-ansible.legacy.dnf Invoked with name=['openvswitch', 'os-net-config', 'ansible-core'] state=present update_cache=True allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:22:25 localhost kernel: SELinux: Converting 490 SID table entries... Nov 27 02:22:25 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:22:25 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:22:25 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:22:25 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:22:25 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:22:25 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:22:25 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:22:25 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=4 res=1 Nov 27 02:22:25 localhost systemd[1]: Started daily update of the root trust anchor for DNSSEC. Nov 27 02:22:28 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:22:28 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:22:28 localhost systemd[1]: Reloading. Nov 27 02:22:28 localhost systemd-rc-local-generator[21123]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:22:29 localhost systemd-sysv-generator[21127]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:22:29 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:22:29 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:22:29 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:22:29 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:22:29 localhost systemd[1]: run-rc5e12939d76044e8b0a5e2d48ad31f92.service: Deactivated successfully. Nov 27 02:22:30 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:22:30 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 02:22:46 localhost python3[21872]: ansible-ansible.legacy.command Invoked with _raw_params=ansible-galaxy collection install ansible.posix#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-000000000015-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:23:03 localhost python3[21892]: ansible-ansible.builtin.file Invoked with path=/etc/os-net-config state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:23:04 localhost python3[21940]: ansible-ansible.legacy.stat Invoked with path=/etc/os-net-config/tripleo_config.yaml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:23:05 localhost python3[21983]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764228184.4196937-332-113575781983140/source dest=/etc/os-net-config/tripleo_config.yaml mode=None follow=False _original_basename=overcloud_net_config.j2 checksum=91bc45728dd9738fc644e3ada9d8642294da29ff backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:23:06 localhost python3[22013]: ansible-community.general.nmcli Invoked with conn_name=ci-private-network state=absent ignore_unsupported_suboptions=False autoconnect=True gw4_ignore_auto=False never_default4=False dns4_ignore_auto=False may_fail4=True gw6_ignore_auto=False dns6_ignore_auto=False mode=balance-rr stp=True priority=128 slavepriority=32 forwarddelay=15 hellotime=2 maxage=20 ageingtime=300 hairpin=False path_cost=100 runner=roundrobin master=None slave_type=None ifname=None type=None ip4=None gw4=None routes4=None routes4_extended=None route_metric4=None routing_rules4=None dns4=None dns4_search=None dns4_options=None method4=None dhcp_client_id=None ip6=None gw6=None dns6=None dns6_search=None dns6_options=None routes6=None routes6_extended=None route_metric6=None method6=None ip_privacy6=None addr_gen_mode6=None miimon=None downdelay=None updelay=None xmit_hash_policy=None arp_interval=None arp_ip_target=None primary=None mtu=None mac=None zone=None runner_hwaddr_policy=None runner_fast_rate=None vlanid=None vlandev=None flags=None ingress=None egress=None vxlan_id=None vxlan_local=None vxlan_remote=None ip_tunnel_dev=None ip_tunnel_local=None ip_tunnel_remote=None ip_tunnel_input_key=NOT_LOGGING_PARAMETER ip_tunnel_output_key=NOT_LOGGING_PARAMETER ssid=None wifi=None wifi_sec=NOT_LOGGING_PARAMETER gsm=None macvlan=None wireguard=None vpn=None transport_mode=None Nov 27 02:23:07 localhost systemd-journald[618]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 89.2 (297 of 333 items), suggesting rotation. Nov 27 02:23:07 localhost systemd-journald[618]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 02:23:07 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:23:07 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:23:07 localhost python3[22034]: ansible-community.general.nmcli Invoked with conn_name=ci-private-network-20 state=absent ignore_unsupported_suboptions=False autoconnect=True gw4_ignore_auto=False never_default4=False dns4_ignore_auto=False may_fail4=True gw6_ignore_auto=False dns6_ignore_auto=False mode=balance-rr stp=True priority=128 slavepriority=32 forwarddelay=15 hellotime=2 maxage=20 ageingtime=300 hairpin=False path_cost=100 runner=roundrobin master=None slave_type=None ifname=None type=None ip4=None gw4=None routes4=None routes4_extended=None route_metric4=None routing_rules4=None dns4=None dns4_search=None dns4_options=None method4=None dhcp_client_id=None ip6=None gw6=None dns6=None dns6_search=None dns6_options=None routes6=None routes6_extended=None route_metric6=None method6=None ip_privacy6=None addr_gen_mode6=None miimon=None downdelay=None updelay=None xmit_hash_policy=None arp_interval=None arp_ip_target=None primary=None mtu=None mac=None zone=None runner_hwaddr_policy=None runner_fast_rate=None vlanid=None vlandev=None flags=None ingress=None egress=None vxlan_id=None vxlan_local=None vxlan_remote=None ip_tunnel_dev=None ip_tunnel_local=None ip_tunnel_remote=None ip_tunnel_input_key=NOT_LOGGING_PARAMETER ip_tunnel_output_key=NOT_LOGGING_PARAMETER ssid=None wifi=None wifi_sec=NOT_LOGGING_PARAMETER gsm=None macvlan=None wireguard=None vpn=None transport_mode=None Nov 27 02:23:07 localhost python3[22054]: ansible-community.general.nmcli Invoked with conn_name=ci-private-network-21 state=absent ignore_unsupported_suboptions=False autoconnect=True gw4_ignore_auto=False never_default4=False dns4_ignore_auto=False may_fail4=True gw6_ignore_auto=False dns6_ignore_auto=False mode=balance-rr stp=True priority=128 slavepriority=32 forwarddelay=15 hellotime=2 maxage=20 ageingtime=300 hairpin=False path_cost=100 runner=roundrobin master=None slave_type=None ifname=None type=None ip4=None gw4=None routes4=None routes4_extended=None route_metric4=None routing_rules4=None dns4=None dns4_search=None dns4_options=None method4=None dhcp_client_id=None ip6=None gw6=None dns6=None dns6_search=None dns6_options=None routes6=None routes6_extended=None route_metric6=None method6=None ip_privacy6=None addr_gen_mode6=None miimon=None downdelay=None updelay=None xmit_hash_policy=None arp_interval=None arp_ip_target=None primary=None mtu=None mac=None zone=None runner_hwaddr_policy=None runner_fast_rate=None vlanid=None vlandev=None flags=None ingress=None egress=None vxlan_id=None vxlan_local=None vxlan_remote=None ip_tunnel_dev=None ip_tunnel_local=None ip_tunnel_remote=None ip_tunnel_input_key=NOT_LOGGING_PARAMETER ip_tunnel_output_key=NOT_LOGGING_PARAMETER ssid=None wifi=None wifi_sec=NOT_LOGGING_PARAMETER gsm=None macvlan=None wireguard=None vpn=None transport_mode=None Nov 27 02:23:07 localhost python3[22074]: ansible-community.general.nmcli Invoked with conn_name=ci-private-network-22 state=absent ignore_unsupported_suboptions=False autoconnect=True gw4_ignore_auto=False never_default4=False dns4_ignore_auto=False may_fail4=True gw6_ignore_auto=False dns6_ignore_auto=False mode=balance-rr stp=True priority=128 slavepriority=32 forwarddelay=15 hellotime=2 maxage=20 ageingtime=300 hairpin=False path_cost=100 runner=roundrobin master=None slave_type=None ifname=None type=None ip4=None gw4=None routes4=None routes4_extended=None route_metric4=None routing_rules4=None dns4=None dns4_search=None dns4_options=None method4=None dhcp_client_id=None ip6=None gw6=None dns6=None dns6_search=None dns6_options=None routes6=None routes6_extended=None route_metric6=None method6=None ip_privacy6=None addr_gen_mode6=None miimon=None downdelay=None updelay=None xmit_hash_policy=None arp_interval=None arp_ip_target=None primary=None mtu=None mac=None zone=None runner_hwaddr_policy=None runner_fast_rate=None vlanid=None vlandev=None flags=None ingress=None egress=None vxlan_id=None vxlan_local=None vxlan_remote=None ip_tunnel_dev=None ip_tunnel_local=None ip_tunnel_remote=None ip_tunnel_input_key=NOT_LOGGING_PARAMETER ip_tunnel_output_key=NOT_LOGGING_PARAMETER ssid=None wifi=None wifi_sec=NOT_LOGGING_PARAMETER gsm=None macvlan=None wireguard=None vpn=None transport_mode=None Nov 27 02:23:08 localhost python3[22094]: ansible-community.general.nmcli Invoked with conn_name=ci-private-network-23 state=absent ignore_unsupported_suboptions=False autoconnect=True gw4_ignore_auto=False never_default4=False dns4_ignore_auto=False may_fail4=True gw6_ignore_auto=False dns6_ignore_auto=False mode=balance-rr stp=True priority=128 slavepriority=32 forwarddelay=15 hellotime=2 maxage=20 ageingtime=300 hairpin=False path_cost=100 runner=roundrobin master=None slave_type=None ifname=None type=None ip4=None gw4=None routes4=None routes4_extended=None route_metric4=None routing_rules4=None dns4=None dns4_search=None dns4_options=None method4=None dhcp_client_id=None ip6=None gw6=None dns6=None dns6_search=None dns6_options=None routes6=None routes6_extended=None route_metric6=None method6=None ip_privacy6=None addr_gen_mode6=None miimon=None downdelay=None updelay=None xmit_hash_policy=None arp_interval=None arp_ip_target=None primary=None mtu=None mac=None zone=None runner_hwaddr_policy=None runner_fast_rate=None vlanid=None vlandev=None flags=None ingress=None egress=None vxlan_id=None vxlan_local=None vxlan_remote=None ip_tunnel_dev=None ip_tunnel_local=None ip_tunnel_remote=None ip_tunnel_input_key=NOT_LOGGING_PARAMETER ip_tunnel_output_key=NOT_LOGGING_PARAMETER ssid=None wifi=None wifi_sec=NOT_LOGGING_PARAMETER gsm=None macvlan=None wireguard=None vpn=None transport_mode=None Nov 27 02:23:11 localhost python3[22114]: ansible-ansible.builtin.systemd Invoked with name=network state=started daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:23:11 localhost systemd[1]: Starting LSB: Bring up/down networking... Nov 27 02:23:11 localhost network[22117]: WARN : [network] You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 02:23:11 localhost network[22128]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 02:23:11 localhost network[22117]: WARN : [network] 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:11 localhost network[22129]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:11 localhost network[22117]: WARN : [network] It is advised to switch to 'NetworkManager' instead for network management. Nov 27 02:23:11 localhost network[22130]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 02:23:11 localhost NetworkManager[5971]: [1764228191.2126] audit: op="connections-reload" pid=22158 uid=0 result="success" Nov 27 02:23:11 localhost network[22117]: Bringing up loopback interface: [ OK ] Nov 27 02:23:11 localhost NetworkManager[5971]: [1764228191.4099] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth0" pid=22246 uid=0 result="success" Nov 27 02:23:11 localhost network[22117]: Bringing up interface eth0: [ OK ] Nov 27 02:23:11 localhost systemd[1]: Started LSB: Bring up/down networking. Nov 27 02:23:11 localhost python3[22287]: ansible-ansible.builtin.systemd Invoked with name=openvswitch state=started daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:23:11 localhost systemd[1]: Starting Open vSwitch Database Unit... Nov 27 02:23:11 localhost chown[22291]: /usr/bin/chown: cannot access '/run/openvswitch': No such file or directory Nov 27 02:23:11 localhost ovs-ctl[22296]: /etc/openvswitch/conf.db does not exist ... (warning). Nov 27 02:23:12 localhost ovs-ctl[22296]: Creating empty database /etc/openvswitch/conf.db [ OK ] Nov 27 02:23:12 localhost ovs-ctl[22296]: Starting ovsdb-server [ OK ] Nov 27 02:23:12 localhost ovs-vsctl[22345]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait -- init -- set Open_vSwitch . db-version=8.5.1 Nov 27 02:23:12 localhost ovs-vsctl[22365]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait set Open_vSwitch . ovs-version=3.3.6-141.el9fdp "external-ids:system-id=\"927ada8e-9e05-4f20-b8c8-d7c6f26f86ee\"" "external-ids:rundir=\"/var/run/openvswitch\"" "system-type=\"rhel\"" "system-version=\"9.2\"" Nov 27 02:23:12 localhost ovs-ctl[22296]: Configuring Open vSwitch system IDs [ OK ] Nov 27 02:23:12 localhost ovs-ctl[22296]: Enabling remote OVSDB managers [ OK ] Nov 27 02:23:12 localhost systemd[1]: Started Open vSwitch Database Unit. Nov 27 02:23:12 localhost ovs-vsctl[22371]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait add Open_vSwitch . external-ids hostname=np0005537446.novalocal Nov 27 02:23:12 localhost systemd[1]: Starting Open vSwitch Delete Transient Ports... Nov 27 02:23:12 localhost systemd[1]: Finished Open vSwitch Delete Transient Ports. Nov 27 02:23:12 localhost systemd[1]: Starting Open vSwitch Forwarding Unit... Nov 27 02:23:12 localhost kernel: openvswitch: Open vSwitch switching datapath Nov 27 02:23:12 localhost ovs-ctl[22415]: Inserting openvswitch module [ OK ] Nov 27 02:23:12 localhost ovs-ctl[22384]: Starting ovs-vswitchd [ OK ] Nov 27 02:23:12 localhost ovs-vsctl[22434]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait add Open_vSwitch . external-ids hostname=np0005537446.novalocal Nov 27 02:23:12 localhost ovs-ctl[22384]: Enabling remote OVSDB managers [ OK ] Nov 27 02:23:12 localhost systemd[1]: Started Open vSwitch Forwarding Unit. Nov 27 02:23:12 localhost systemd[1]: Starting Open vSwitch... Nov 27 02:23:12 localhost systemd[1]: Finished Open vSwitch. Nov 27 02:23:43 localhost python3[22452]: ansible-ansible.legacy.command Invoked with _raw_params=os-net-config -c /etc/os-net-config/tripleo_config.yaml#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-00000000001a-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.0932] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22611 uid=0 result="success" Nov 27 02:23:44 localhost ifup[22612]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:44 localhost ifup[22613]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:44 localhost ifup[22614]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.1257] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22620 uid=0 result="success" Nov 27 02:23:44 localhost ovs-vsctl[22622]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --may-exist add-br br-ex -- set bridge br-ex other-config:mac-table-size=50000 -- set bridge br-ex other-config:hwaddr=fa:16:3e:cd:69:4f -- set bridge br-ex fail_mode=standalone -- del-controller br-ex Nov 27 02:23:44 localhost kernel: device ovs-system entered promiscuous mode Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.1819] manager: (ovs-system): new Generic device (/org/freedesktop/NetworkManager/Devices/4) Nov 27 02:23:44 localhost kernel: Timeout policy base is empty Nov 27 02:23:44 localhost kernel: Failed to associated timeout policy `ovs_test_tp' Nov 27 02:23:44 localhost systemd-udevd[22623]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:44 localhost kernel: device br-ex entered promiscuous mode Nov 27 02:23:44 localhost systemd-udevd[22638]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.2277] manager: (br-ex): new Generic device (/org/freedesktop/NetworkManager/Devices/5) Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.2558] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22649 uid=0 result="success" Nov 27 02:23:44 localhost NetworkManager[5971]: [1764228224.2783] device (br-ex): carrier: link connected Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.3361] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22678 uid=0 result="success" Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.3878] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22693 uid=0 result="success" Nov 27 02:23:47 localhost NET[22718]: /etc/sysconfig/network-scripts/ifup-post : updated /etc/resolv.conf Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.4815] device (eth1): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'managed') Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.4965] dhcp4 (eth1): canceled DHCP transaction Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.4965] dhcp4 (eth1): activation: beginning transaction (timeout in 45 seconds) Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.4965] dhcp4 (eth1): state changed no lease Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.5018] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth1" pid=22727 uid=0 result="success" Nov 27 02:23:47 localhost ifup[22728]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:47 localhost ifup[22729]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:47 localhost ifup[22730]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:47 localhost systemd[1]: Starting Network Manager Script Dispatcher Service... Nov 27 02:23:47 localhost systemd[1]: Started Network Manager Script Dispatcher Service. Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.5458] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth1" pid=22744 uid=0 result="success" Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.5964] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth1" pid=22755 uid=0 result="success" Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.6037] device (eth1): carrier: link connected Nov 27 02:23:47 localhost NetworkManager[5971]: [1764228227.6263] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth1" pid=22764 uid=0 result="success" Nov 27 02:23:47 localhost ipv6_wait_tentative[22776]: Waiting for interface eth1 IPv6 address(es) to leave the 'tentative' state Nov 27 02:23:48 localhost ipv6_wait_tentative[22781]: Waiting for interface eth1 IPv6 address(es) to leave the 'tentative' state Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.6976] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-eth1" pid=22790 uid=0 result="success" Nov 27 02:23:49 localhost ovs-vsctl[22805]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex eth1 -- add-port br-ex eth1 Nov 27 02:23:49 localhost kernel: device eth1 entered promiscuous mode Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.8105] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22813 uid=0 result="success" Nov 27 02:23:49 localhost ifup[22814]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:49 localhost ifup[22815]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:49 localhost ifup[22816]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.8408] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-br-ex" pid=22822 uid=0 result="success" Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.8818] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=22832 uid=0 result="success" Nov 27 02:23:49 localhost ifup[22833]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:49 localhost ifup[22834]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:49 localhost ifup[22835]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.9131] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=22841 uid=0 result="success" Nov 27 02:23:49 localhost ovs-vsctl[22844]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan21 -- add-port br-ex vlan21 tag=21 -- set Interface vlan21 type=internal Nov 27 02:23:49 localhost kernel: device vlan21 entered promiscuous mode Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.9544] manager: (vlan21): new Generic device (/org/freedesktop/NetworkManager/Devices/6) Nov 27 02:23:49 localhost systemd-udevd[22846]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.9766] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=22855 uid=0 result="success" Nov 27 02:23:49 localhost NetworkManager[5971]: [1764228229.9979] device (vlan21): carrier: link connected Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.0549] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=22884 uid=0 result="success" Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.1048] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=22899 uid=0 result="success" Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.1696] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=22920 uid=0 result="success" Nov 27 02:23:53 localhost ifup[22921]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:53 localhost ifup[22922]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:53 localhost ifup[22923]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.2025] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=22929 uid=0 result="success" Nov 27 02:23:53 localhost ovs-vsctl[22932]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan20 -- add-port br-ex vlan20 tag=20 -- set Interface vlan20 type=internal Nov 27 02:23:53 localhost kernel: device vlan20 entered promiscuous mode Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.2461] manager: (vlan20): new Generic device (/org/freedesktop/NetworkManager/Devices/7) Nov 27 02:23:53 localhost systemd-udevd[22935]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.2749] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=22944 uid=0 result="success" Nov 27 02:23:53 localhost NetworkManager[5971]: [1764228233.2976] device (vlan20): carrier: link connected Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.3412] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=22974 uid=0 result="success" Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.3831] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=22989 uid=0 result="success" Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.4360] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23010 uid=0 result="success" Nov 27 02:23:56 localhost ifup[23011]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:56 localhost ifup[23012]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:56 localhost ifup[23013]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.4639] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23019 uid=0 result="success" Nov 27 02:23:56 localhost ovs-vsctl[23022]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan23 -- add-port br-ex vlan23 tag=23 -- set Interface vlan23 type=internal Nov 27 02:23:56 localhost systemd-udevd[23024]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.5023] manager: (vlan23): new Generic device (/org/freedesktop/NetworkManager/Devices/8) Nov 27 02:23:56 localhost kernel: device vlan23 entered promiscuous mode Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.5275] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23034 uid=0 result="success" Nov 27 02:23:56 localhost NetworkManager[5971]: [1764228236.5465] device (vlan23): carrier: link connected Nov 27 02:23:57 localhost systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.5963] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23064 uid=0 result="success" Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.6445] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23079 uid=0 result="success" Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.7045] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23100 uid=0 result="success" Nov 27 02:23:59 localhost ifup[23101]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:23:59 localhost ifup[23102]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:23:59 localhost ifup[23103]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.7376] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23109 uid=0 result="success" Nov 27 02:23:59 localhost ovs-vsctl[23112]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan44 -- add-port br-ex vlan44 tag=44 -- set Interface vlan44 type=internal Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.7782] manager: (vlan44): new Generic device (/org/freedesktop/NetworkManager/Devices/9) Nov 27 02:23:59 localhost kernel: device vlan44 entered promiscuous mode Nov 27 02:23:59 localhost systemd-udevd[23114]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.8064] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23124 uid=0 result="success" Nov 27 02:23:59 localhost NetworkManager[5971]: [1764228239.8289] device (vlan44): carrier: link connected Nov 27 02:24:02 localhost NetworkManager[5971]: [1764228242.8862] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23154 uid=0 result="success" Nov 27 02:24:02 localhost NetworkManager[5971]: [1764228242.9348] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23169 uid=0 result="success" Nov 27 02:24:02 localhost NetworkManager[5971]: [1764228242.9946] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23190 uid=0 result="success" Nov 27 02:24:02 localhost ifup[23191]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:03 localhost ifup[23192]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:03 localhost ifup[23193]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:03 localhost NetworkManager[5971]: [1764228243.0268] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23199 uid=0 result="success" Nov 27 02:24:03 localhost ovs-vsctl[23202]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan22 -- add-port br-ex vlan22 tag=22 -- set Interface vlan22 type=internal Nov 27 02:24:03 localhost systemd-udevd[23204]: Network interface NamePolicy= disabled on kernel command line. Nov 27 02:24:03 localhost kernel: device vlan22 entered promiscuous mode Nov 27 02:24:03 localhost NetworkManager[5971]: [1764228243.0681] manager: (vlan22): new Generic device (/org/freedesktop/NetworkManager/Devices/10) Nov 27 02:24:03 localhost NetworkManager[5971]: [1764228243.0960] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23214 uid=0 result="success" Nov 27 02:24:03 localhost NetworkManager[5971]: [1764228243.1170] device (vlan22): carrier: link connected Nov 27 02:24:06 localhost NetworkManager[5971]: [1764228246.1690] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23244 uid=0 result="success" Nov 27 02:24:06 localhost NetworkManager[5971]: [1764228246.2163] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23259 uid=0 result="success" Nov 27 02:24:06 localhost NetworkManager[5971]: [1764228246.2746] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23280 uid=0 result="success" Nov 27 02:24:06 localhost ifup[23281]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:06 localhost ifup[23282]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:06 localhost ifup[23283]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:06 localhost NetworkManager[5971]: [1764228246.3065] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23289 uid=0 result="success" Nov 27 02:24:06 localhost ovs-vsctl[23292]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan44 -- add-port br-ex vlan44 tag=44 -- set Interface vlan44 type=internal Nov 27 02:24:06 localhost NetworkManager[5971]: [1764228246.3660] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23299 uid=0 result="success" Nov 27 02:24:07 localhost NetworkManager[5971]: [1764228247.4238] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23326 uid=0 result="success" Nov 27 02:24:07 localhost NetworkManager[5971]: [1764228247.4706] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan44" pid=23341 uid=0 result="success" Nov 27 02:24:07 localhost NetworkManager[5971]: [1764228247.5291] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=23362 uid=0 result="success" Nov 27 02:24:07 localhost ifup[23363]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:07 localhost ifup[23364]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:07 localhost ifup[23365]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:07 localhost NetworkManager[5971]: [1764228247.5603] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=23371 uid=0 result="success" Nov 27 02:24:07 localhost ovs-vsctl[23374]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan20 -- add-port br-ex vlan20 tag=20 -- set Interface vlan20 type=internal Nov 27 02:24:07 localhost NetworkManager[5971]: [1764228247.6169] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=23381 uid=0 result="success" Nov 27 02:24:08 localhost NetworkManager[5971]: [1764228248.6755] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=23410 uid=0 result="success" Nov 27 02:24:08 localhost NetworkManager[5971]: [1764228248.7219] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan20" pid=23425 uid=0 result="success" Nov 27 02:24:08 localhost NetworkManager[5971]: [1764228248.7811] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=23446 uid=0 result="success" Nov 27 02:24:08 localhost ifup[23447]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:08 localhost ifup[23448]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:08 localhost ifup[23449]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:08 localhost NetworkManager[5971]: [1764228248.8132] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=23455 uid=0 result="success" Nov 27 02:24:08 localhost ovs-vsctl[23458]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan21 -- add-port br-ex vlan21 tag=21 -- set Interface vlan21 type=internal Nov 27 02:24:08 localhost NetworkManager[5971]: [1764228248.9137] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=23465 uid=0 result="success" Nov 27 02:24:09 localhost NetworkManager[5971]: [1764228249.9753] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=23493 uid=0 result="success" Nov 27 02:24:10 localhost NetworkManager[5971]: [1764228250.0245] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan21" pid=23508 uid=0 result="success" Nov 27 02:24:10 localhost NetworkManager[5971]: [1764228250.0849] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23529 uid=0 result="success" Nov 27 02:24:10 localhost ifup[23530]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:10 localhost ifup[23531]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:10 localhost ifup[23532]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:10 localhost NetworkManager[5971]: [1764228250.1189] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23538 uid=0 result="success" Nov 27 02:24:10 localhost ovs-vsctl[23541]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan23 -- add-port br-ex vlan23 tag=23 -- set Interface vlan23 type=internal Nov 27 02:24:10 localhost NetworkManager[5971]: [1764228250.2260] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23548 uid=0 result="success" Nov 27 02:24:11 localhost NetworkManager[5971]: [1764228251.2783] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23576 uid=0 result="success" Nov 27 02:24:11 localhost NetworkManager[5971]: [1764228251.3192] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan23" pid=23591 uid=0 result="success" Nov 27 02:24:11 localhost NetworkManager[5971]: [1764228251.3782] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23612 uid=0 result="success" Nov 27 02:24:11 localhost ifup[23613]: You are using 'ifup' script provided by 'network-scripts', which are now deprecated. Nov 27 02:24:11 localhost ifup[23614]: 'network-scripts' will be removed from distribution in near future. Nov 27 02:24:11 localhost ifup[23615]: It is advised to switch to 'NetworkManager' instead - it provides 'ifup/ifdown' scripts as well. Nov 27 02:24:11 localhost NetworkManager[5971]: [1764228251.4092] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23621 uid=0 result="success" Nov 27 02:24:11 localhost ovs-vsctl[23624]: ovs|00001|vsctl|INFO|Called as ovs-vsctl -t 10 -- --if-exists del-port br-ex vlan22 -- add-port br-ex vlan22 tag=22 -- set Interface vlan22 type=internal Nov 27 02:24:11 localhost NetworkManager[5971]: [1764228251.4932] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23631 uid=0 result="success" Nov 27 02:24:12 localhost NetworkManager[5971]: [1764228252.5477] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23659 uid=0 result="success" Nov 27 02:24:12 localhost NetworkManager[5971]: [1764228252.5938] audit: op="connections-load" args="/etc/sysconfig/network-scripts/ifcfg-vlan22" pid=23674 uid=0 result="success" Nov 27 02:24:37 localhost python3[23706]: ansible-ansible.legacy.command Invoked with _raw_params=ip a#012ping -c 2 -W 2 192.168.122.10#012ping -c 2 -W 2 192.168.122.11#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-00000000001b-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:24:42 localhost python3[23725]: ansible-ansible.posix.authorized_key Invoked with user=zuul key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:24:42 localhost python3[23741]: ansible-ansible.posix.authorized_key Invoked with user=root key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:24:43 localhost python3[23755]: ansible-ansible.posix.authorized_key Invoked with user=zuul key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:24:44 localhost python3[23771]: ansible-ansible.posix.authorized_key Invoked with user=root key=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey manage_dir=True state=present exclusive=False validate_certs=True follow=False path=None key_options=None comment=None Nov 27 02:24:44 localhost python3[23785]: ansible-ansible.builtin.slurp Invoked with path=/etc/hostname src=/etc/hostname Nov 27 02:24:45 localhost python3[23800]: ansible-ansible.legacy.command Invoked with _raw_params=hostname="np0005537446.novalocal"#012hostname_str_array=(${hostname//./ })#012echo ${hostname_str_array[0]} > /home/zuul/ansible_hostname#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-000000000022-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:24:46 localhost python3[23820]: ansible-ansible.legacy.command Invoked with _raw_params=hostname=$(cat /home/zuul/ansible_hostname)#012hostnamectl hostname "$hostname.localdomain"#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-0cb1-928f-000000000023-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:24:46 localhost systemd[1]: Starting Hostname Service... Nov 27 02:24:46 localhost systemd[1]: Started Hostname Service. Nov 27 02:24:46 localhost systemd-hostnamed[23824]: Hostname set to (static) Nov 27 02:24:46 localhost NetworkManager[5971]: [1764228286.7440] hostname: static hostname changed from "np0005537446.novalocal" to "np0005537446.localdomain" Nov 27 02:24:46 localhost systemd[1]: Starting Network Manager Script Dispatcher Service... Nov 27 02:24:46 localhost systemd[1]: Started Network Manager Script Dispatcher Service. Nov 27 02:24:47 localhost systemd[1]: session-10.scope: Deactivated successfully. Nov 27 02:24:47 localhost systemd[1]: session-10.scope: Consumed 1min 45.739s CPU time. Nov 27 02:24:47 localhost systemd-logind[761]: Session 10 logged out. Waiting for processes to exit. Nov 27 02:24:47 localhost systemd-logind[761]: Removed session 10. Nov 27 02:24:52 localhost sshd[23835]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:24:52 localhost systemd-logind[761]: New session 11 of user zuul. Nov 27 02:24:52 localhost systemd[1]: Started Session 11 of User zuul. Nov 27 02:24:52 localhost python3[23852]: ansible-ansible.builtin.slurp Invoked with path=/home/zuul/ansible_hostname src=/home/zuul/ansible_hostname Nov 27 02:24:53 localhost systemd[1]: session-11.scope: Deactivated successfully. Nov 27 02:24:53 localhost systemd-logind[761]: Session 11 logged out. Waiting for processes to exit. Nov 27 02:24:53 localhost systemd-logind[761]: Removed session 11. Nov 27 02:24:56 localhost systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. Nov 27 02:25:16 localhost systemd[1]: systemd-hostnamed.service: Deactivated successfully. Nov 27 02:25:45 localhost sshd[23857]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:25:46 localhost systemd-logind[761]: New session 12 of user zuul. Nov 27 02:25:46 localhost systemd[1]: Started Session 12 of User zuul. Nov 27 02:25:46 localhost python3[23876]: ansible-ansible.legacy.dnf Invoked with name=['lvm2', 'jq'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:25:50 localhost systemd[1]: Reloading. Nov 27 02:25:50 localhost systemd-rc-local-generator[23914]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:25:50 localhost systemd-sysv-generator[23920]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:25:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:25:50 localhost systemd[1]: Listening on Device-mapper event daemon FIFOs. Nov 27 02:25:50 localhost systemd[1]: Reloading. Nov 27 02:25:51 localhost systemd-sysv-generator[23958]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:25:51 localhost systemd-rc-local-generator[23954]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:25:51 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:25:51 localhost systemd[1]: Starting Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling... Nov 27 02:25:51 localhost systemd[1]: Finished Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling. Nov 27 02:25:51 localhost systemd[1]: Reloading. Nov 27 02:25:51 localhost systemd-rc-local-generator[23993]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:25:51 localhost systemd-sysv-generator[24000]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:25:51 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:25:51 localhost systemd[1]: Listening on LVM2 poll daemon socket. Nov 27 02:25:51 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:25:51 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:25:51 localhost systemd[1]: Reloading. Nov 27 02:25:51 localhost systemd-sysv-generator[24062]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:25:51 localhost systemd-rc-local-generator[24057]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:25:51 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:25:52 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:25:52 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:25:52 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:25:52 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:25:52 localhost systemd[1]: run-rc100ba4d734b428884c395331c0ef43b.service: Deactivated successfully. Nov 27 02:25:52 localhost systemd[1]: run-re39179e275914c7abdf500b825fd5ba0.service: Deactivated successfully. Nov 27 02:26:19 localhost sshd[24646]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:26:52 localhost systemd[1]: session-12.scope: Deactivated successfully. Nov 27 02:26:52 localhost systemd[1]: session-12.scope: Consumed 4.923s CPU time. Nov 27 02:26:52 localhost systemd-logind[761]: Session 12 logged out. Waiting for processes to exit. Nov 27 02:26:52 localhost systemd-logind[761]: Removed session 12. Nov 27 02:43:03 localhost sshd[24656]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:43:03 localhost systemd-logind[761]: New session 13 of user zuul. Nov 27 02:43:03 localhost systemd[1]: Started Session 13 of User zuul. Nov 27 02:43:04 localhost python3[24704]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 02:43:06 localhost python3[24791]: ansible-ansible.builtin.dnf Invoked with name=['util-linux', 'lvm2', 'jq', 'podman'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:43:09 localhost python3[24808]: ansible-ansible.builtin.stat Invoked with path=/dev/loop3 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:43:09 localhost python3[24824]: ansible-ansible.legacy.command Invoked with _raw_params=dd if=/dev/zero of=/var/lib/ceph-osd-0.img bs=1 count=0 seek=7G#012losetup /dev/loop3 /var/lib/ceph-osd-0.img#012lsblk _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:09 localhost kernel: loop: module loaded Nov 27 02:43:09 localhost kernel: loop3: detected capacity change from 0 to 14680064 Nov 27 02:43:10 localhost python3[24849]: ansible-ansible.legacy.command Invoked with _raw_params=pvcreate /dev/loop3#012vgcreate ceph_vg0 /dev/loop3#012lvcreate -n ceph_lv0 -l +100%FREE ceph_vg0#012lvs _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:10 localhost lvm[24852]: PV /dev/loop3 not used. Nov 27 02:43:10 localhost lvm[24861]: PV /dev/loop3 online, VG ceph_vg0 is complete. Nov 27 02:43:10 localhost systemd[1]: Started /usr/sbin/lvm vgchange -aay --autoactivation event ceph_vg0. Nov 27 02:43:10 localhost lvm[24863]: 1 logical volume(s) in volume group "ceph_vg0" now active Nov 27 02:43:10 localhost systemd[1]: lvm-activate-ceph_vg0.service: Deactivated successfully. Nov 27 02:43:11 localhost python3[24911]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/ceph-osd-losetup-0.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:43:11 localhost python3[24954]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764229390.941343-54352-55607469857470/source dest=/etc/systemd/system/ceph-osd-losetup-0.service mode=0644 force=True follow=False _original_basename=ceph-osd-losetup.service.j2 checksum=427b1db064a970126b729b07acf99fa7d0eecb9c backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:12 localhost python3[24984]: ansible-ansible.builtin.systemd Invoked with state=started enabled=True name=ceph-osd-losetup-0.service daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:43:13 localhost systemd[1]: Reloading. Nov 27 02:43:13 localhost systemd-sysv-generator[25016]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:43:13 localhost systemd-rc-local-generator[25011]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:43:13 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:43:13 localhost systemd[1]: Starting Ceph OSD losetup... Nov 27 02:43:13 localhost bash[25025]: /dev/loop3: [64516]:8399529 (/var/lib/ceph-osd-0.img) Nov 27 02:43:13 localhost systemd[1]: Finished Ceph OSD losetup. Nov 27 02:43:13 localhost lvm[25027]: PV /dev/loop3 online, VG ceph_vg0 is complete. Nov 27 02:43:13 localhost lvm[25027]: VG ceph_vg0 finished Nov 27 02:43:14 localhost python3[25044]: ansible-ansible.builtin.dnf Invoked with name=['util-linux', 'lvm2', 'jq', 'podman'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:43:17 localhost python3[25061]: ansible-ansible.builtin.stat Invoked with path=/dev/loop4 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:43:18 localhost python3[25077]: ansible-ansible.legacy.command Invoked with _raw_params=dd if=/dev/zero of=/var/lib/ceph-osd-1.img bs=1 count=0 seek=7G#012losetup /dev/loop4 /var/lib/ceph-osd-1.img#012lsblk _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:18 localhost kernel: loop4: detected capacity change from 0 to 14680064 Nov 27 02:43:18 localhost python3[25099]: ansible-ansible.legacy.command Invoked with _raw_params=pvcreate /dev/loop4#012vgcreate ceph_vg1 /dev/loop4#012lvcreate -n ceph_lv1 -l +100%FREE ceph_vg1#012lvs _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:18 localhost lvm[25102]: PV /dev/loop4 not used. Nov 27 02:43:18 localhost lvm[25104]: PV /dev/loop4 online, VG ceph_vg1 is complete. Nov 27 02:43:18 localhost systemd[1]: Started /usr/sbin/lvm vgchange -aay --autoactivation event ceph_vg1. Nov 27 02:43:18 localhost lvm[25111]: 1 logical volume(s) in volume group "ceph_vg1" now active Nov 27 02:43:18 localhost lvm[25115]: PV /dev/loop4 online, VG ceph_vg1 is complete. Nov 27 02:43:18 localhost lvm[25115]: VG ceph_vg1 finished Nov 27 02:43:18 localhost systemd[1]: lvm-activate-ceph_vg1.service: Deactivated successfully. Nov 27 02:43:19 localhost python3[25163]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/ceph-osd-losetup-1.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:43:19 localhost python3[25206]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764229398.9939554-54504-272695420022117/source dest=/etc/systemd/system/ceph-osd-losetup-1.service mode=0644 force=True follow=False _original_basename=ceph-osd-losetup.service.j2 checksum=19612168ea279db4171b94ee1f8625de1ec44b58 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:20 localhost python3[25236]: ansible-ansible.builtin.systemd Invoked with state=started enabled=True name=ceph-osd-losetup-1.service daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:43:20 localhost systemd[1]: Reloading. Nov 27 02:43:20 localhost systemd-sysv-generator[25266]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:43:20 localhost systemd-rc-local-generator[25261]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:43:20 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:43:20 localhost systemd[1]: Starting Ceph OSD losetup... Nov 27 02:43:20 localhost bash[25276]: /dev/loop4: [64516]:8400144 (/var/lib/ceph-osd-1.img) Nov 27 02:43:20 localhost systemd[1]: Finished Ceph OSD losetup. Nov 27 02:43:20 localhost lvm[25277]: PV /dev/loop4 online, VG ceph_vg1 is complete. Nov 27 02:43:20 localhost lvm[25277]: VG ceph_vg1 finished Nov 27 02:43:30 localhost python3[25322]: ansible-ansible.legacy.setup Invoked with gather_subset=['!all', 'min'] gather_timeout=45 filter=[] fact_path=/etc/ansible/facts.d Nov 27 02:43:31 localhost python3[25342]: ansible-hostname Invoked with name=np0005537446.localdomain use=None Nov 27 02:43:31 localhost systemd[1]: Starting Hostname Service... Nov 27 02:43:32 localhost systemd[1]: Started Hostname Service. Nov 27 02:43:34 localhost python3[25365]: ansible-tempfile Invoked with state=file suffix=tmphosts prefix=ansible. path=None Nov 27 02:43:34 localhost python3[25413]: ansible-ansible.legacy.copy Invoked with remote_src=True src=/etc/hosts dest=/tmp/ansible.964bw_jltmphosts mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:35 localhost python3[25443]: ansible-blockinfile Invoked with state=absent path=/tmp/ansible.964bw_jltmphosts block= marker=# {mark} marker_begin=HEAT_HOSTS_START - Do not edit manually within this section! marker_end=HEAT_HOSTS_END create=False backup=False unsafe_writes=False insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:35 localhost python3[25459]: ansible-blockinfile Invoked with create=True path=/tmp/ansible.964bw_jltmphosts insertbefore=BOF block=192.168.122.106 np0005537444.localdomain np0005537444#012192.168.122.106 np0005537444.ctlplane.localdomain np0005537444.ctlplane#012192.168.122.107 np0005537445.localdomain np0005537445#012192.168.122.107 np0005537445.ctlplane.localdomain np0005537445.ctlplane#012192.168.122.108 np0005537446.localdomain np0005537446#012192.168.122.108 np0005537446.ctlplane.localdomain np0005537446.ctlplane#012192.168.122.103 np0005537441.localdomain np0005537441#012192.168.122.103 np0005537441.ctlplane.localdomain np0005537441.ctlplane#012192.168.122.104 np0005537442.localdomain np0005537442#012192.168.122.104 np0005537442.ctlplane.localdomain np0005537442.ctlplane#012192.168.122.105 np0005537443.localdomain np0005537443#012192.168.122.105 np0005537443.ctlplane.localdomain np0005537443.ctlplane#012#012192.168.122.100 undercloud.ctlplane.localdomain undercloud.ctlplane#012 marker=# {mark} marker_begin=START_HOST_ENTRIES_FOR_STACK: overcloud marker_end=END_HOST_ENTRIES_FOR_STACK: overcloud state=present backup=False unsafe_writes=False insertafter=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:36 localhost python3[25476]: ansible-ansible.legacy.command Invoked with _raw_params=cp "/tmp/ansible.964bw_jltmphosts" "/etc/hosts" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:36 localhost python3[25493]: ansible-file Invoked with path=/tmp/ansible.964bw_jltmphosts state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:39 localhost python3[25509]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-active ntpd.service || systemctl is-enabled ntpd.service _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:39 localhost python3[25527]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:43:43 localhost python3[25576]: ansible-ansible.legacy.stat Invoked with path=/etc/chrony.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:43:44 localhost python3[25621]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764229423.145921-55460-247475166963726/source dest=/etc/chrony.conf owner=root group=root mode=420 follow=False _original_basename=chrony.conf.j2 checksum=4fd4fbbb2de00c70a54478b7feb8ef8adf6a3362 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:45 localhost python3[25651]: ansible-ansible.legacy.systemd Invoked with enabled=True name=chronyd state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:43:47 localhost python3[25669]: ansible-ansible.legacy.systemd Invoked with name=chronyd state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:43:47 localhost chronyd[767]: chronyd exiting Nov 27 02:43:47 localhost systemd[1]: Stopping NTP client/server... Nov 27 02:43:47 localhost systemd[1]: chronyd.service: Deactivated successfully. Nov 27 02:43:47 localhost systemd[1]: Stopped NTP client/server. Nov 27 02:43:47 localhost systemd[1]: chronyd.service: Consumed 96ms CPU time, read 1.9M from disk, written 0B to disk. Nov 27 02:43:47 localhost systemd[1]: Starting NTP client/server... Nov 27 02:43:47 localhost chronyd[25676]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Nov 27 02:43:47 localhost chronyd[25676]: Frequency -26.220 +/- 0.193 ppm read from /var/lib/chrony/drift Nov 27 02:43:47 localhost chronyd[25676]: Loaded seccomp filter (level 2) Nov 27 02:43:47 localhost systemd[1]: Started NTP client/server. Nov 27 02:43:49 localhost python3[25725]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/chrony-online.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:43:49 localhost python3[25768]: ansible-ansible.legacy.copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764229428.75885-55626-107259935260594/source dest=/etc/systemd/system/chrony-online.service _original_basename=chrony-online.service follow=False checksum=d4d85e046d61f558ac7ec8178c6d529d893e81e1 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:43:50 localhost python3[25798]: ansible-systemd Invoked with state=started name=chrony-online.service enabled=True daemon-reload=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:43:50 localhost systemd[1]: Reloading. Nov 27 02:43:50 localhost systemd-sysv-generator[25824]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:43:50 localhost systemd-rc-local-generator[25821]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:43:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:43:50 localhost systemd[1]: Reloading. Nov 27 02:43:50 localhost systemd-rc-local-generator[25861]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:43:50 localhost systemd-sysv-generator[25868]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:43:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:43:50 localhost systemd[1]: Starting chronyd online sources service... Nov 27 02:43:50 localhost chronyc[25875]: 200 OK Nov 27 02:43:50 localhost systemd[1]: chrony-online.service: Deactivated successfully. Nov 27 02:43:50 localhost systemd[1]: Finished chronyd online sources service. Nov 27 02:43:51 localhost python3[25891]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc makestep _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:51 localhost chronyd[25676]: System clock was stepped by 0.000000 seconds Nov 27 02:43:51 localhost python3[25908]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 30 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:43:52 localhost chronyd[25676]: Selected source 167.160.187.179 (pool.ntp.org) Nov 27 02:44:02 localhost systemd[1]: systemd-hostnamed.service: Deactivated successfully. Nov 27 02:44:02 localhost python3[25925]: ansible-timezone Invoked with name=UTC hwclock=None Nov 27 02:44:02 localhost systemd[1]: Starting Time & Date Service... Nov 27 02:44:02 localhost systemd[1]: Started Time & Date Service. Nov 27 02:44:03 localhost python3[25948]: ansible-ansible.legacy.systemd Invoked with name=chronyd state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:44:03 localhost systemd[1]: Stopping NTP client/server... Nov 27 02:44:03 localhost chronyd[25676]: chronyd exiting Nov 27 02:44:03 localhost systemd[1]: chronyd.service: Deactivated successfully. Nov 27 02:44:03 localhost systemd[1]: Stopped NTP client/server. Nov 27 02:44:03 localhost systemd[1]: Starting NTP client/server... Nov 27 02:44:03 localhost chronyd[25955]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Nov 27 02:44:03 localhost chronyd[25955]: Frequency -26.220 +/- 0.196 ppm read from /var/lib/chrony/drift Nov 27 02:44:03 localhost chronyd[25955]: Loaded seccomp filter (level 2) Nov 27 02:44:03 localhost systemd[1]: Started NTP client/server. Nov 27 02:44:07 localhost chronyd[25955]: Selected source 23.159.16.194 (pool.ntp.org) Nov 27 02:44:32 localhost systemd[1]: systemd-timedated.service: Deactivated successfully. Nov 27 02:44:49 localhost sshd[26154]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:12 localhost sshd[26155]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:12 localhost systemd[1]: Created slice User Slice of UID 1002. Nov 27 02:46:12 localhost systemd[1]: Starting User Runtime Directory /run/user/1002... Nov 27 02:46:12 localhost systemd-logind[761]: New session 14 of user ceph-admin. Nov 27 02:46:12 localhost systemd[1]: Finished User Runtime Directory /run/user/1002. Nov 27 02:46:12 localhost systemd[1]: Starting User Manager for UID 1002... Nov 27 02:46:12 localhost sshd[26172]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:12 localhost systemd[26159]: Queued start job for default target Main User Target. Nov 27 02:46:12 localhost systemd[26159]: Created slice User Application Slice. Nov 27 02:46:12 localhost systemd[26159]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 02:46:12 localhost systemd[26159]: Started Daily Cleanup of User's Temporary Directories. Nov 27 02:46:12 localhost systemd[26159]: Reached target Paths. Nov 27 02:46:12 localhost systemd[26159]: Reached target Timers. Nov 27 02:46:12 localhost systemd[26159]: Starting D-Bus User Message Bus Socket... Nov 27 02:46:12 localhost systemd[26159]: Starting Create User's Volatile Files and Directories... Nov 27 02:46:12 localhost systemd[26159]: Listening on D-Bus User Message Bus Socket. Nov 27 02:46:12 localhost systemd[26159]: Reached target Sockets. Nov 27 02:46:12 localhost systemd[26159]: Finished Create User's Volatile Files and Directories. Nov 27 02:46:12 localhost systemd[26159]: Reached target Basic System. Nov 27 02:46:12 localhost systemd[26159]: Reached target Main User Target. Nov 27 02:46:12 localhost systemd[26159]: Startup finished in 115ms. Nov 27 02:46:12 localhost systemd[1]: Started User Manager for UID 1002. Nov 27 02:46:12 localhost systemd[1]: Started Session 14 of User ceph-admin. Nov 27 02:46:12 localhost systemd-logind[761]: New session 16 of user ceph-admin. Nov 27 02:46:12 localhost systemd[1]: Started Session 16 of User ceph-admin. Nov 27 02:46:13 localhost sshd[26194]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:13 localhost systemd-logind[761]: New session 17 of user ceph-admin. Nov 27 02:46:13 localhost systemd[1]: Started Session 17 of User ceph-admin. Nov 27 02:46:13 localhost sshd[26213]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:13 localhost systemd-logind[761]: New session 18 of user ceph-admin. Nov 27 02:46:13 localhost systemd[1]: Started Session 18 of User ceph-admin. Nov 27 02:46:13 localhost sshd[26232]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:13 localhost systemd-logind[761]: New session 19 of user ceph-admin. Nov 27 02:46:13 localhost systemd[1]: Started Session 19 of User ceph-admin. Nov 27 02:46:14 localhost sshd[26251]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:14 localhost systemd-logind[761]: New session 20 of user ceph-admin. Nov 27 02:46:14 localhost systemd[1]: Started Session 20 of User ceph-admin. Nov 27 02:46:14 localhost sshd[26270]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:14 localhost systemd-logind[761]: New session 21 of user ceph-admin. Nov 27 02:46:14 localhost systemd[1]: Started Session 21 of User ceph-admin. Nov 27 02:46:14 localhost sshd[26289]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:15 localhost systemd-logind[761]: New session 22 of user ceph-admin. Nov 27 02:46:15 localhost systemd[1]: Started Session 22 of User ceph-admin. Nov 27 02:46:15 localhost sshd[26308]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:15 localhost systemd-logind[761]: New session 23 of user ceph-admin. Nov 27 02:46:15 localhost systemd[1]: Started Session 23 of User ceph-admin. Nov 27 02:46:15 localhost sshd[26327]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:15 localhost systemd-logind[761]: New session 24 of user ceph-admin. Nov 27 02:46:15 localhost systemd[1]: Started Session 24 of User ceph-admin. Nov 27 02:46:16 localhost sshd[26344]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:16 localhost systemd-logind[761]: New session 25 of user ceph-admin. Nov 27 02:46:16 localhost systemd[1]: Started Session 25 of User ceph-admin. Nov 27 02:46:16 localhost sshd[26363]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:46:16 localhost systemd-logind[761]: New session 26 of user ceph-admin. Nov 27 02:46:16 localhost systemd[1]: Started Session 26 of User ceph-admin. Nov 27 02:46:17 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:32 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:32 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:33 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:33 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:33 localhost systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 26578 (sysctl) Nov 27 02:46:33 localhost systemd[1]: Mounting Arbitrary Executable File Formats File System... Nov 27 02:46:33 localhost systemd[1]: Mounted Arbitrary Executable File Formats File System. Nov 27 02:46:34 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:34 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:46:38 localhost kernel: VFS: idmapped mount is not enabled. Nov 27 02:46:59 localhost podman[26716]: Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:59.053940598 +0000 UTC m=+24.112998499 container create 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, distribution-scope=public, io.openshift.expose-services=, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, RELEASE=main, ceph=True, io.buildah.version=1.33.12, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, release=553, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:34.985413231 +0000 UTC m=+0.044471152 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:46:59 localhost systemd[1]: Created slice Slice /machine. Nov 27 02:46:59 localhost systemd[1]: Started libpod-conmon-59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d.scope. Nov 27 02:46:59 localhost systemd[1]: Started libcrun container. Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:59.159851471 +0000 UTC m=+24.218909402 container init 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, version=7, ceph=True, release=553, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, com.redhat.component=rhceph-container, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, vcs-type=git, GIT_CLEAN=True, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, maintainer=Guillaume Abrioux ) Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:59.169930701 +0000 UTC m=+24.228988632 container start 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, GIT_BRANCH=main, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., vcs-type=git, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, ceph=True) Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:59.170191879 +0000 UTC m=+24.229249870 container attach 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, GIT_BRANCH=main, release=553, io.buildah.version=1.33.12, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, vcs-type=git, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph) Nov 27 02:46:59 localhost keen_hopper[26910]: 167 167 Nov 27 02:46:59 localhost systemd[1]: libpod-59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d.scope: Deactivated successfully. Nov 27 02:46:59 localhost podman[26716]: 2025-11-27 07:46:59.175699099 +0000 UTC m=+24.234756990 container died 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, architecture=x86_64, description=Red Hat Ceph Storage 7, release=553, io.buildah.version=1.33.12, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, ceph=True, io.openshift.expose-services=, name=rhceph, vcs-type=git, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=) Nov 27 02:46:59 localhost podman[26915]: 2025-11-27 07:46:59.310573025 +0000 UTC m=+0.125536979 container remove 59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=keen_hopper, vcs-type=git, GIT_BRANCH=main, io.openshift.expose-services=, version=7, io.buildah.version=1.33.12, RELEASE=main, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, release=553, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:46:59 localhost systemd[1]: libpod-conmon-59b7d7018e04271e81d8bb1d3d445c3f000354d721520bf37f3d37e6c9ba262d.scope: Deactivated successfully. Nov 27 02:46:59 localhost podman[27027]: Nov 27 02:46:59 localhost podman[27027]: 2025-11-27 07:46:59.560067231 +0000 UTC m=+0.055942605 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:00 localhost systemd[1]: tmp-crun.ZxdjDg.mount: Deactivated successfully. Nov 27 02:47:00 localhost systemd[1]: var-lib-containers-storage-overlay-7848f01c422da9cf4b3ab346c9e4b1cf053986576a904c5f9c828d5b13b5ea1e-merged.mount: Deactivated successfully. Nov 27 02:47:03 localhost podman[27027]: 2025-11-27 07:47:03.001609076 +0000 UTC m=+3.497484450 container create 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, RELEASE=main, version=7, build-date=2025-09-24T08:57:55, ceph=True, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, name=rhceph, io.buildah.version=1.33.12, release=553, vendor=Red Hat, Inc., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:03 localhost systemd[1]: Started libpod-conmon-489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934.scope. Nov 27 02:47:03 localhost systemd[1]: Started libcrun container. Nov 27 02:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/46b6c11d1aa0a2eefd08f364ad542419f007aa8e58906a9b5c853751ff21782a/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/46b6c11d1aa0a2eefd08f364ad542419f007aa8e58906a9b5c853751ff21782a/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:03 localhost podman[27027]: 2025-11-27 07:47:03.168896659 +0000 UTC m=+3.664772033 container init 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, version=7, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, vcs-type=git, name=rhceph, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, ceph=True, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., GIT_BRANCH=main, release=553, distribution-scope=public, build-date=2025-09-24T08:57:55, RELEASE=main) Nov 27 02:47:03 localhost podman[27027]: 2025-11-27 07:47:03.181571859 +0000 UTC m=+3.677447233 container start 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, com.redhat.component=rhceph-container, architecture=x86_64, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, RELEASE=main, ceph=True, io.openshift.expose-services=, version=7, release=553, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 02:47:03 localhost podman[27027]: 2025-11-27 07:47:03.18452627 +0000 UTC m=+3.680401644 container attach 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, distribution-scope=public, RELEASE=main, architecture=x86_64, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, version=7, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, ceph=True, vcs-type=git, GIT_BRANCH=main, vendor=Red Hat, Inc., name=rhceph) Nov 27 02:47:03 localhost bold_almeida[27209]: [ Nov 27 02:47:03 localhost bold_almeida[27209]: { Nov 27 02:47:03 localhost bold_almeida[27209]: "available": false, Nov 27 02:47:03 localhost bold_almeida[27209]: "ceph_device": false, Nov 27 02:47:03 localhost bold_almeida[27209]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 02:47:03 localhost bold_almeida[27209]: "lsm_data": {}, Nov 27 02:47:03 localhost bold_almeida[27209]: "lvs": [], Nov 27 02:47:03 localhost bold_almeida[27209]: "path": "/dev/sr0", Nov 27 02:47:03 localhost bold_almeida[27209]: "rejected_reasons": [ Nov 27 02:47:03 localhost bold_almeida[27209]: "Insufficient space (<5GB)", Nov 27 02:47:03 localhost bold_almeida[27209]: "Has a FileSystem" Nov 27 02:47:03 localhost bold_almeida[27209]: ], Nov 27 02:47:03 localhost bold_almeida[27209]: "sys_api": { Nov 27 02:47:03 localhost bold_almeida[27209]: "actuators": null, Nov 27 02:47:03 localhost bold_almeida[27209]: "device_nodes": "sr0", Nov 27 02:47:03 localhost bold_almeida[27209]: "human_readable_size": "482.00 KB", Nov 27 02:47:03 localhost bold_almeida[27209]: "id_bus": "ata", Nov 27 02:47:03 localhost bold_almeida[27209]: "model": "QEMU DVD-ROM", Nov 27 02:47:03 localhost bold_almeida[27209]: "nr_requests": "2", Nov 27 02:47:03 localhost bold_almeida[27209]: "partitions": {}, Nov 27 02:47:03 localhost bold_almeida[27209]: "path": "/dev/sr0", Nov 27 02:47:03 localhost bold_almeida[27209]: "removable": "1", Nov 27 02:47:03 localhost bold_almeida[27209]: "rev": "2.5+", Nov 27 02:47:03 localhost bold_almeida[27209]: "ro": "0", Nov 27 02:47:03 localhost bold_almeida[27209]: "rotational": "1", Nov 27 02:47:03 localhost bold_almeida[27209]: "sas_address": "", Nov 27 02:47:03 localhost bold_almeida[27209]: "sas_device_handle": "", Nov 27 02:47:03 localhost bold_almeida[27209]: "scheduler_mode": "mq-deadline", Nov 27 02:47:03 localhost bold_almeida[27209]: "sectors": 0, Nov 27 02:47:03 localhost bold_almeida[27209]: "sectorsize": "2048", Nov 27 02:47:03 localhost bold_almeida[27209]: "size": 493568.0, Nov 27 02:47:03 localhost bold_almeida[27209]: "support_discard": "0", Nov 27 02:47:03 localhost bold_almeida[27209]: "type": "disk", Nov 27 02:47:03 localhost bold_almeida[27209]: "vendor": "QEMU" Nov 27 02:47:03 localhost bold_almeida[27209]: } Nov 27 02:47:03 localhost bold_almeida[27209]: } Nov 27 02:47:03 localhost bold_almeida[27209]: ] Nov 27 02:47:03 localhost systemd[1]: libpod-489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934.scope: Deactivated successfully. Nov 27 02:47:03 localhost podman[27027]: 2025-11-27 07:47:03.918869854 +0000 UTC m=+4.414745298 container died 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, architecture=x86_64, name=rhceph, RELEASE=main, GIT_CLEAN=True, io.buildah.version=1.33.12, distribution-scope=public, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, ceph=True, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, release=553) Nov 27 02:47:03 localhost systemd[1]: tmp-crun.HBsaJN.mount: Deactivated successfully. Nov 27 02:47:04 localhost podman[28514]: 2025-11-27 07:47:04.003728928 +0000 UTC m=+0.075962931 container remove 489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_almeida, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_BRANCH=main, io.buildah.version=1.33.12, release=553, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, architecture=x86_64, RELEASE=main, vendor=Red Hat, Inc., distribution-scope=public, com.redhat.component=rhceph-container) Nov 27 02:47:04 localhost systemd[1]: var-lib-containers-storage-overlay-46b6c11d1aa0a2eefd08f364ad542419f007aa8e58906a9b5c853751ff21782a-merged.mount: Deactivated successfully. Nov 27 02:47:04 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:04 localhost systemd[1]: libpod-conmon-489a4c34dd59545f73d268b66eb90a4dab24172fde115d1c9aebbae542143934.scope: Deactivated successfully. Nov 27 02:47:04 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:04 localhost systemd[1]: systemd-coredump.socket: Deactivated successfully. Nov 27 02:47:04 localhost systemd[1]: Closed Process Core Dump Socket. Nov 27 02:47:04 localhost systemd[1]: Stopping Process Core Dump Socket... Nov 27 02:47:04 localhost systemd[1]: Listening on Process Core Dump Socket. Nov 27 02:47:04 localhost systemd[1]: Reloading. Nov 27 02:47:04 localhost systemd-sysv-generator[28596]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:04 localhost systemd-rc-local-generator[28593]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:04 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:04 localhost systemd[1]: Reloading. Nov 27 02:47:04 localhost systemd-sysv-generator[28640]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:04 localhost systemd-rc-local-generator[28634]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:04 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:29 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:30 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:30 localhost podman[28717]: Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.164517336 +0000 UTC m=+0.077790714 container create 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, distribution-scope=public, GIT_BRANCH=main, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, GIT_CLEAN=True, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , release=553) Nov 27 02:47:30 localhost systemd[1]: Started libpod-conmon-16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c.scope. Nov 27 02:47:30 localhost systemd[1]: Started libcrun container. Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.229567988 +0000 UTC m=+0.142841366 container init 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, io.openshift.tags=rhceph ceph, release=553, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, name=rhceph, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, GIT_CLEAN=True, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, CEPH_POINT_RELEASE=, ceph=True, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, distribution-scope=public, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, version=7, vcs-type=git) Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.132542259 +0000 UTC m=+0.045815637 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.238966771 +0000 UTC m=+0.152240159 container start 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, version=7, RELEASE=main, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, release=553, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., ceph=True, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , vcs-type=git, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, distribution-scope=public) Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.239222377 +0000 UTC m=+0.152495805 container attach 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, io.buildah.version=1.33.12, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, maintainer=Guillaume Abrioux , RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, release=553) Nov 27 02:47:30 localhost busy_grothendieck[28733]: 167 167 Nov 27 02:47:30 localhost systemd[1]: libpod-16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c.scope: Deactivated successfully. Nov 27 02:47:30 localhost podman[28717]: 2025-11-27 07:47:30.243313873 +0000 UTC m=+0.156587251 container died 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, GIT_CLEAN=True, vendor=Red Hat, Inc., GIT_BRANCH=main, io.openshift.tags=rhceph ceph, distribution-scope=public, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12) Nov 27 02:47:30 localhost podman[28738]: 2025-11-27 07:47:30.326797842 +0000 UTC m=+0.075778747 container remove 16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_grothendieck, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, name=rhceph, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., RELEASE=main, ceph=True, maintainer=Guillaume Abrioux , release=553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, architecture=x86_64, GIT_BRANCH=main, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git) Nov 27 02:47:30 localhost systemd[1]: libpod-conmon-16dcef45413504b1f4136e0ed4290ff0a216b8b9c10746c11a6938ff7de0d60c.scope: Deactivated successfully. Nov 27 02:47:30 localhost systemd[1]: Reloading. Nov 27 02:47:30 localhost systemd-sysv-generator[28780]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:30 localhost systemd-rc-local-generator[28777]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:30 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:30 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:30 localhost systemd[1]: Reloading. Nov 27 02:47:30 localhost systemd-sysv-generator[28816]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:30 localhost systemd-rc-local-generator[28812]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:30 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:30 localhost systemd[1]: Reached target All Ceph clusters and services. Nov 27 02:47:30 localhost systemd[1]: Reloading. Nov 27 02:47:31 localhost systemd-sysv-generator[28855]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:31 localhost systemd-rc-local-generator[28851]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:31 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:31 localhost systemd[1]: Reached target Ceph cluster e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 02:47:31 localhost systemd[1]: Reloading. Nov 27 02:47:31 localhost systemd-rc-local-generator[28894]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:31 localhost systemd-sysv-generator[28897]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:31 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:31 localhost systemd[1]: Reloading. Nov 27 02:47:31 localhost systemd-rc-local-generator[28935]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:31 localhost systemd-sysv-generator[28940]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:31 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:31 localhost systemd[1]: Created slice Slice /system/ceph-e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 02:47:31 localhost systemd[1]: Reached target System Time Set. Nov 27 02:47:31 localhost systemd[1]: Reached target System Time Synchronized. Nov 27 02:47:31 localhost systemd[1]: Starting Ceph crash.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 02:47:31 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:31 localhost systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. Nov 27 02:47:31 localhost podman[28999]: Nov 27 02:47:32 localhost podman[28999]: 2025-11-27 07:47:32.005081273 +0000 UTC m=+0.073778989 container create 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, name=rhceph, description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, architecture=x86_64, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, vcs-type=git, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=) Nov 27 02:47:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/14796c8c2ad77d885100c28b1b0f65a4225e9634b7302a2fa5fccca738e21467/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/14796c8c2ad77d885100c28b1b0f65a4225e9634b7302a2fa5fccca738e21467/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:32 localhost podman[28999]: 2025-11-27 07:47:31.977136971 +0000 UTC m=+0.045834697 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/14796c8c2ad77d885100c28b1b0f65a4225e9634b7302a2fa5fccca738e21467/merged/etc/ceph/ceph.client.crash.np0005537446.keyring supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:32 localhost podman[28999]: 2025-11-27 07:47:32.096685974 +0000 UTC m=+0.165383690 container init 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, name=rhceph, ceph=True, description=Red Hat Ceph Storage 7, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, GIT_CLEAN=True, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , architecture=x86_64, release=553, com.redhat.component=rhceph-container, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., io.buildah.version=1.33.12) Nov 27 02:47:32 localhost podman[28999]: 2025-11-27 07:47:32.107688385 +0000 UTC m=+0.176386101 container start 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, maintainer=Guillaume Abrioux , vcs-type=git, ceph=True, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, name=rhceph, GIT_CLEAN=True, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, CEPH_POINT_RELEASE=, distribution-scope=public, version=7, com.redhat.component=rhceph-container, GIT_BRANCH=main, vendor=Red Hat, Inc., architecture=x86_64) Nov 27 02:47:32 localhost bash[28999]: 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 Nov 27 02:47:32 localhost systemd[1]: Started Ceph crash.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: INFO:ceph-crash:pinging cluster to exercise our key Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.269+0000 7f1c2cf62640 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin: (2) No such file or directory Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.269+0000 7f1c2cf62640 -1 AuthRegistry(0x7f1c28068980) no keyring found at /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin, disabling cephx Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.271+0000 7f1c2cf62640 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin: (2) No such file or directory Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.271+0000 7f1c2cf62640 -1 AuthRegistry(0x7f1c2cf61000) no keyring found at /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin, disabling cephx Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.281+0000 7f1c26d76640 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.282+0000 7f1c26575640 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.284+0000 7f1c25d74640 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: 2025-11-27T07:47:32.284+0000 7f1c2cf62640 -1 monclient: authenticate NOTE: no keyring found; disabled cephx authentication Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: [errno 13] RADOS permission denied (error connecting to the cluster) Nov 27 02:47:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446[29013]: INFO:ceph-crash:monitoring path /var/lib/ceph/crash, delay 600s Nov 27 02:47:35 localhost podman[29100]: Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.667289749 +0000 UTC m=+0.067250015 container create 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , vcs-type=git, GIT_CLEAN=True, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, name=rhceph, com.redhat.component=rhceph-container, GIT_BRANCH=main, vendor=Red Hat, Inc., io.buildah.version=1.33.12, release=553, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, version=7) Nov 27 02:47:35 localhost systemd[1]: Started libpod-conmon-792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa.scope. Nov 27 02:47:35 localhost systemd[1]: Started libcrun container. Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.737106243 +0000 UTC m=+0.137066509 container init 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, vcs-type=git, GIT_BRANCH=main, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, release=553, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, RELEASE=main, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.640447612 +0000 UTC m=+0.040407878 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.746574257 +0000 UTC m=+0.146534523 container start 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, maintainer=Guillaume Abrioux , version=7, io.openshift.expose-services=, distribution-scope=public, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, release=553, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.746846214 +0000 UTC m=+0.146806520 container attach 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, io.buildah.version=1.33.12, ceph=True, RELEASE=main, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, architecture=x86_64, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., vcs-type=git, version=7, io.openshift.expose-services=, description=Red Hat Ceph Storage 7) Nov 27 02:47:35 localhost angry_liskov[29115]: 167 167 Nov 27 02:47:35 localhost systemd[1]: libpod-792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa.scope: Deactivated successfully. Nov 27 02:47:35 localhost podman[29100]: 2025-11-27 07:47:35.750589332 +0000 UTC m=+0.150549648 container died 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, version=7, io.openshift.tags=rhceph ceph, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, io.openshift.expose-services=, RELEASE=main, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vcs-type=git, ceph=True, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, distribution-scope=public) Nov 27 02:47:35 localhost systemd[1]: var-lib-containers-storage-overlay-4a2707a6e1e6d02a157b3ed663e460b817f966c5ea1fd3eafd49d592e86b3523-merged.mount: Deactivated successfully. Nov 27 02:47:35 localhost podman[29120]: 2025-11-27 07:47:35.842530711 +0000 UTC m=+0.081611755 container remove 792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_liskov, distribution-scope=public, version=7, architecture=x86_64, io.openshift.expose-services=, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, release=553, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, GIT_CLEAN=True) Nov 27 02:47:35 localhost systemd[1]: libpod-conmon-792c23116ae893196eef7845935f797a95454870cbd848cdcaca8eecd8e59bfa.scope: Deactivated successfully. Nov 27 02:47:36 localhost podman[29140]: Nov 27 02:47:36 localhost podman[29140]: 2025-11-27 07:47:36.043838682 +0000 UTC m=+0.070505622 container create 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, release=553, maintainer=Guillaume Abrioux , io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, vcs-type=git, distribution-scope=public, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, GIT_CLEAN=True, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, description=Red Hat Ceph Storage 7, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 02:47:36 localhost systemd[1]: Started libpod-conmon-78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5.scope. Nov 27 02:47:36 localhost systemd[1]: Started libcrun container. Nov 27 02:47:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:36 localhost podman[29140]: 2025-11-27 07:47:36.017346874 +0000 UTC m=+0.044013824 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f/merged/var/lib/ceph/bootstrap-osd/ceph.keyring supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:36 localhost podman[29140]: 2025-11-27 07:47:36.163798205 +0000 UTC m=+0.190465135 container init 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, version=7, GIT_CLEAN=True, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, ceph=True, vendor=Red Hat, Inc., GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, maintainer=Guillaume Abrioux , io.openshift.expose-services=, com.redhat.component=rhceph-container, name=rhceph, description=Red Hat Ceph Storage 7) Nov 27 02:47:36 localhost podman[29140]: 2025-11-27 07:47:36.173511624 +0000 UTC m=+0.200178564 container start 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , GIT_BRANCH=main, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, distribution-scope=public, GIT_CLEAN=True, release=553, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 02:47:36 localhost podman[29140]: 2025-11-27 07:47:36.173919904 +0000 UTC m=+0.200586834 container attach 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, io.openshift.expose-services=, distribution-scope=public, release=553, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, vcs-type=git, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 02:47:36 localhost boring_shannon[29155]: --> passed data devices: 0 physical, 2 LVM Nov 27 02:47:36 localhost boring_shannon[29155]: --> relative data size: 1.0 Nov 27 02:47:36 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-authtool --gen-print-key Nov 27 02:47:36 localhost boring_shannon[29155]: Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 042da33d-4846-4886-b930-3adda459b2df Nov 27 02:47:37 localhost lvm[29209]: PV /dev/loop3 online, VG ceph_vg0 is complete. Nov 27 02:47:37 localhost lvm[29209]: VG ceph_vg0 finished Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-authtool --gen-print-key Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/chown -h ceph:ceph /dev/ceph_vg0/ceph_lv0 Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-0 Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/ln -s /dev/ceph_vg0/ceph_lv0 /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap Nov 27 02:47:37 localhost boring_shannon[29155]: stderr: got monmap epoch 3 Nov 27 02:47:37 localhost boring_shannon[29155]: --> Creating keyring file for osd.0 Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ Nov 27 02:47:37 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osdspec-affinity default_drive_group --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid 042da33d-4846-4886-b930-3adda459b2df --setuser ceph --setgroup ceph Nov 27 02:47:40 localhost boring_shannon[29155]: stderr: 2025-11-27T07:47:37.839+0000 7f960f99aa80 -1 bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to decode label at offset 102: void bluestore_bdev_label_t::decode(ceph::buffer::v15_2_0::list::const_iterator&) decode past end of struct encoding: Malformed input [buffer:3] Nov 27 02:47:40 localhost boring_shannon[29155]: stderr: 2025-11-27T07:47:37.839+0000 7f960f99aa80 -1 bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid Nov 27 02:47:40 localhost boring_shannon[29155]: --> ceph-volume lvm prepare successful for: ceph_vg0/ceph_lv0 Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph_vg0/ceph_lv0 --path /var/lib/ceph/osd/ceph-0 --no-mon-config Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/ln -snf /dev/ceph_vg0/ceph_lv0 /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-0 Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:40 localhost boring_shannon[29155]: --> ceph-volume lvm activate successful for osd ID: 0 Nov 27 02:47:40 localhost boring_shannon[29155]: --> ceph-volume lvm create successful for: ceph_vg0/ceph_lv0 Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-authtool --gen-print-key Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new c197799b-1f02-4d5b-a065-671596d4ef72 Nov 27 02:47:40 localhost lvm[30154]: PV /dev/loop4 online, VG ceph_vg1 is complete. Nov 27 02:47:40 localhost lvm[30154]: VG ceph_vg1 finished Nov 27 02:47:40 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-authtool --gen-print-key Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-3 Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/chown -h ceph:ceph /dev/ceph_vg1/ceph_lv1 Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/ln -s /dev/ceph_vg1/ceph_lv1 /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-3/activate.monmap Nov 27 02:47:41 localhost boring_shannon[29155]: stderr: got monmap epoch 3 Nov 27 02:47:41 localhost boring_shannon[29155]: --> Creating keyring file for osd.3 Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/keyring Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/ Nov 27 02:47:41 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 3 --monmap /var/lib/ceph/osd/ceph-3/activate.monmap --keyfile - --osdspec-affinity default_drive_group --osd-data /var/lib/ceph/osd/ceph-3/ --osd-uuid c197799b-1f02-4d5b-a065-671596d4ef72 --setuser ceph --setgroup ceph Nov 27 02:47:43 localhost boring_shannon[29155]: stderr: 2025-11-27T07:47:41.564+0000 7f3fc5718a80 -1 bluestore(/var/lib/ceph/osd/ceph-3//block) _read_bdev_label unable to decode label at offset 102: void bluestore_bdev_label_t::decode(ceph::buffer::v15_2_0::list::const_iterator&) decode past end of struct encoding: Malformed input [buffer:3] Nov 27 02:47:43 localhost boring_shannon[29155]: stderr: 2025-11-27T07:47:41.565+0000 7f3fc5718a80 -1 bluestore(/var/lib/ceph/osd/ceph-3/) _read_fsid unparsable uuid Nov 27 02:47:43 localhost boring_shannon[29155]: --> ceph-volume lvm prepare successful for: ceph_vg1/ceph_lv1 Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph_vg1/ceph_lv1 --path /var/lib/ceph/osd/ceph-3 --no-mon-config Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/ln -snf /dev/ceph_vg1/ceph_lv1 /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 Nov 27 02:47:44 localhost boring_shannon[29155]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:44 localhost boring_shannon[29155]: --> ceph-volume lvm activate successful for osd ID: 3 Nov 27 02:47:44 localhost boring_shannon[29155]: --> ceph-volume lvm create successful for: ceph_vg1/ceph_lv1 Nov 27 02:47:44 localhost systemd[1]: libpod-78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5.scope: Deactivated successfully. Nov 27 02:47:44 localhost systemd[1]: libpod-78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5.scope: Consumed 3.629s CPU time. Nov 27 02:47:44 localhost podman[29140]: 2025-11-27 07:47:44.169829439 +0000 UTC m=+8.196496419 container died 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, release=553, maintainer=Guillaume Abrioux , version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, distribution-scope=public, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, RELEASE=main, architecture=x86_64, name=rhceph, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 02:47:44 localhost systemd[1]: var-lib-containers-storage-overlay-7a48171dd9e96698a260437ef84321640f68679140199223266341f52ba1ac9f-merged.mount: Deactivated successfully. Nov 27 02:47:44 localhost podman[31071]: 2025-11-27 07:47:44.24839092 +0000 UTC m=+0.066132508 container remove 78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=boring_shannon, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_CLEAN=True, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, architecture=x86_64, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, name=rhceph, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., ceph=True, version=7) Nov 27 02:47:44 localhost systemd[1]: libpod-conmon-78a4b878e47a3582330f380ef4eaf688e47b4c2357a9055147d82e4e63aad4c5.scope: Deactivated successfully. Nov 27 02:47:44 localhost podman[31155]: Nov 27 02:47:44 localhost podman[31155]: 2025-11-27 07:47:44.971777113 +0000 UTC m=+0.075744766 container create 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, maintainer=Guillaume Abrioux , release=553, distribution-scope=public, GIT_BRANCH=main, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, vcs-type=git, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, io.buildah.version=1.33.12, version=7, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 02:47:45 localhost systemd[1]: Started libpod-conmon-73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff.scope. Nov 27 02:47:45 localhost systemd[1]: Started libcrun container. Nov 27 02:47:45 localhost podman[31155]: 2025-11-27 07:47:45.039301833 +0000 UTC m=+0.143269436 container init 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, io.openshift.expose-services=, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, release=553, GIT_BRANCH=main, RELEASE=main, version=7, CEPH_POINT_RELEASE=) Nov 27 02:47:45 localhost podman[31155]: 2025-11-27 07:47:44.941085515 +0000 UTC m=+0.045053158 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:45 localhost podman[31155]: 2025-11-27 07:47:45.054550845 +0000 UTC m=+0.158518478 container start 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, distribution-scope=public, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, release=553, RELEASE=main, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main) Nov 27 02:47:45 localhost podman[31155]: 2025-11-27 07:47:45.054898473 +0000 UTC m=+0.158866076 container attach 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, version=7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, vcs-type=git, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., RELEASE=main, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, io.buildah.version=1.33.12, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, distribution-scope=public, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 02:47:45 localhost objective_joliot[31171]: 167 167 Nov 27 02:47:45 localhost systemd[1]: libpod-73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff.scope: Deactivated successfully. Nov 27 02:47:45 localhost podman[31155]: 2025-11-27 07:47:45.06026203 +0000 UTC m=+0.164229653 container died 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, name=rhceph, GIT_BRANCH=main, architecture=x86_64, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vcs-type=git, version=7) Nov 27 02:47:45 localhost podman[31176]: 2025-11-27 07:47:45.152527007 +0000 UTC m=+0.078335638 container remove 73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=objective_joliot, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, architecture=x86_64, ceph=True, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-type=git, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , GIT_CLEAN=True) Nov 27 02:47:45 localhost systemd[1]: libpod-conmon-73a4bb05de6bf7f858dc8f65c9c84d317a4fd9ae74229fb985166ab3fb1223ff.scope: Deactivated successfully. Nov 27 02:47:45 localhost systemd[1]: tmp-crun.IPuWDC.mount: Deactivated successfully. Nov 27 02:47:45 localhost systemd[1]: var-lib-containers-storage-overlay-7d913be9e0537f27360bef63b6dc836082b64cbd00342a723a25352482f9c7eb-merged.mount: Deactivated successfully. Nov 27 02:47:45 localhost podman[31195]: Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.356558241 +0000 UTC m=+0.068798471 container create cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, release=553, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, distribution-scope=public, ceph=True, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, name=rhceph, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, version=7, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.openshift.expose-services=, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55) Nov 27 02:47:45 localhost systemd[1]: Started libpod-conmon-cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548.scope. Nov 27 02:47:45 localhost systemd[1]: Started libcrun container. Nov 27 02:47:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3c70866789fc51bdb954a93771b7386c3adcf541a9d98c04c8f5f31207580bd2/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3c70866789fc51bdb954a93771b7386c3adcf541a9d98c04c8f5f31207580bd2/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.330127455 +0000 UTC m=+0.042367715 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3c70866789fc51bdb954a93771b7386c3adcf541a9d98c04c8f5f31207580bd2/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.445325585 +0000 UTC m=+0.157565815 container init cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, name=rhceph, distribution-scope=public, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.454079102 +0000 UTC m=+0.166319352 container start cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, GIT_CLEAN=True, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, ceph=True, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.455001694 +0000 UTC m=+0.167241944 container attach cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, release=553, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, RELEASE=main, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 02:47:45 localhost adoring_benz[31211]: { Nov 27 02:47:45 localhost adoring_benz[31211]: "0": [ Nov 27 02:47:45 localhost adoring_benz[31211]: { Nov 27 02:47:45 localhost adoring_benz[31211]: "devices": [ Nov 27 02:47:45 localhost adoring_benz[31211]: "/dev/loop3" Nov 27 02:47:45 localhost adoring_benz[31211]: ], Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_name": "ceph_lv0", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_path": "/dev/ceph_vg0/ceph_lv0", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_size": "7511998464", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_tags": "ceph.block_device=/dev/ceph_vg0/ceph_lv0,ceph.block_uuid=z1aFYQ-rwZ2-wFiJ-QUte-o2q4-fTHd-mh7dbP,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=e83f3b0c-4090-52df-95d4-ad9be8516692,ceph.cluster_name=ceph,ceph.crush_device_class=,ceph.encrypted=0,ceph.osd_fsid=042da33d-4846-4886-b930-3adda459b2df,ceph.osd_id=0,ceph.osdspec_affinity=default_drive_group,ceph.type=block,ceph.vdo=0", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_uuid": "z1aFYQ-rwZ2-wFiJ-QUte-o2q4-fTHd-mh7dbP", Nov 27 02:47:45 localhost adoring_benz[31211]: "name": "ceph_lv0", Nov 27 02:47:45 localhost adoring_benz[31211]: "path": "/dev/ceph_vg0/ceph_lv0", Nov 27 02:47:45 localhost adoring_benz[31211]: "tags": { Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.block_device": "/dev/ceph_vg0/ceph_lv0", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.block_uuid": "z1aFYQ-rwZ2-wFiJ-QUte-o2q4-fTHd-mh7dbP", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cephx_lockbox_secret": "", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cluster_fsid": "e83f3b0c-4090-52df-95d4-ad9be8516692", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cluster_name": "ceph", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.crush_device_class": "", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.encrypted": "0", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osd_fsid": "042da33d-4846-4886-b930-3adda459b2df", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osd_id": "0", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osdspec_affinity": "default_drive_group", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.type": "block", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.vdo": "0" Nov 27 02:47:45 localhost adoring_benz[31211]: }, Nov 27 02:47:45 localhost adoring_benz[31211]: "type": "block", Nov 27 02:47:45 localhost adoring_benz[31211]: "vg_name": "ceph_vg0" Nov 27 02:47:45 localhost adoring_benz[31211]: } Nov 27 02:47:45 localhost adoring_benz[31211]: ], Nov 27 02:47:45 localhost adoring_benz[31211]: "3": [ Nov 27 02:47:45 localhost adoring_benz[31211]: { Nov 27 02:47:45 localhost adoring_benz[31211]: "devices": [ Nov 27 02:47:45 localhost adoring_benz[31211]: "/dev/loop4" Nov 27 02:47:45 localhost adoring_benz[31211]: ], Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_name": "ceph_lv1", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_path": "/dev/ceph_vg1/ceph_lv1", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_size": "7511998464", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_tags": "ceph.block_device=/dev/ceph_vg1/ceph_lv1,ceph.block_uuid=D03Rjd-YT9L-4fwB-LkRp-A1Iy-nd6B-DpYptJ,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=e83f3b0c-4090-52df-95d4-ad9be8516692,ceph.cluster_name=ceph,ceph.crush_device_class=,ceph.encrypted=0,ceph.osd_fsid=c197799b-1f02-4d5b-a065-671596d4ef72,ceph.osd_id=3,ceph.osdspec_affinity=default_drive_group,ceph.type=block,ceph.vdo=0", Nov 27 02:47:45 localhost adoring_benz[31211]: "lv_uuid": "D03Rjd-YT9L-4fwB-LkRp-A1Iy-nd6B-DpYptJ", Nov 27 02:47:45 localhost adoring_benz[31211]: "name": "ceph_lv1", Nov 27 02:47:45 localhost adoring_benz[31211]: "path": "/dev/ceph_vg1/ceph_lv1", Nov 27 02:47:45 localhost adoring_benz[31211]: "tags": { Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.block_device": "/dev/ceph_vg1/ceph_lv1", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.block_uuid": "D03Rjd-YT9L-4fwB-LkRp-A1Iy-nd6B-DpYptJ", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cephx_lockbox_secret": "", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cluster_fsid": "e83f3b0c-4090-52df-95d4-ad9be8516692", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.cluster_name": "ceph", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.crush_device_class": "", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.encrypted": "0", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osd_fsid": "c197799b-1f02-4d5b-a065-671596d4ef72", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osd_id": "3", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.osdspec_affinity": "default_drive_group", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.type": "block", Nov 27 02:47:45 localhost adoring_benz[31211]: "ceph.vdo": "0" Nov 27 02:47:45 localhost adoring_benz[31211]: }, Nov 27 02:47:45 localhost adoring_benz[31211]: "type": "block", Nov 27 02:47:45 localhost adoring_benz[31211]: "vg_name": "ceph_vg1" Nov 27 02:47:45 localhost adoring_benz[31211]: } Nov 27 02:47:45 localhost adoring_benz[31211]: ] Nov 27 02:47:45 localhost adoring_benz[31211]: } Nov 27 02:47:45 localhost systemd[1]: libpod-cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548.scope: Deactivated successfully. Nov 27 02:47:45 localhost podman[31195]: 2025-11-27 07:47:45.808898561 +0000 UTC m=+0.521138841 container died cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, release=553, io.openshift.tags=rhceph ceph, name=rhceph, ceph=True, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, distribution-scope=public, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, com.redhat.component=rhceph-container, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=) Nov 27 02:47:45 localhost podman[31220]: 2025-11-27 07:47:45.902613631 +0000 UTC m=+0.080062478 container remove cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=adoring_benz, vcs-type=git, RELEASE=main, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, GIT_BRANCH=main, io.openshift.expose-services=, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, ceph=True, io.buildah.version=1.33.12, release=553) Nov 27 02:47:45 localhost systemd[1]: libpod-conmon-cfa22b12a14042ba20b8793aae458379440aa3549aecb71e1e975251a9b77548.scope: Deactivated successfully. Nov 27 02:47:46 localhost systemd[1]: var-lib-containers-storage-overlay-3c70866789fc51bdb954a93771b7386c3adcf541a9d98c04c8f5f31207580bd2-merged.mount: Deactivated successfully. Nov 27 02:47:46 localhost podman[31307]: Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.677121626 +0000 UTC m=+0.069422047 container create 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, maintainer=Guillaume Abrioux , io.openshift.expose-services=, release=553, distribution-scope=public, GIT_CLEAN=True, vcs-type=git, name=rhceph, architecture=x86_64, RELEASE=main, build-date=2025-09-24T08:57:55, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., GIT_BRANCH=main, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:46 localhost systemd[1]: Started libpod-conmon-083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b.scope. Nov 27 02:47:46 localhost systemd[1]: Started libcrun container. Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.741377898 +0000 UTC m=+0.133678319 container init 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, GIT_BRANCH=main, com.redhat.component=rhceph-container, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, RELEASE=main, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, release=553, description=Red Hat Ceph Storage 7, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , GIT_CLEAN=True, vcs-type=git, version=7) Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.650093735 +0000 UTC m=+0.042394156 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.750427633 +0000 UTC m=+0.142728054 container start 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, release=553, com.redhat.component=rhceph-container, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, maintainer=Guillaume Abrioux , distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., GIT_BRANCH=main, name=rhceph, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, architecture=x86_64, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.750641968 +0000 UTC m=+0.142942389 container attach 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, architecture=x86_64, release=553, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, RELEASE=main, io.openshift.expose-services=, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, vcs-type=git, io.openshift.tags=rhceph ceph, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:46 localhost determined_chaum[31322]: 167 167 Nov 27 02:47:46 localhost systemd[1]: libpod-083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b.scope: Deactivated successfully. Nov 27 02:47:46 localhost podman[31307]: 2025-11-27 07:47:46.754193351 +0000 UTC m=+0.146493822 container died 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, GIT_BRANCH=main, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, RELEASE=main, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, name=rhceph, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, com.redhat.component=rhceph-container, release=553) Nov 27 02:47:46 localhost podman[31327]: 2025-11-27 07:47:46.84695903 +0000 UTC m=+0.079021324 container remove 083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_chaum, io.openshift.expose-services=, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, distribution-scope=public, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, GIT_BRANCH=main, version=7, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:46 localhost systemd[1]: libpod-conmon-083057e45c3dca12a3279f93d5e03d63e5b3906d66dd6d573b6807c3a723838b.scope: Deactivated successfully. Nov 27 02:47:47 localhost podman[31356]: Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.169090053 +0000 UTC m=+0.074281140 container create 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, name=rhceph, release=553, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, build-date=2025-09-24T08:57:55, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., GIT_CLEAN=True, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git) Nov 27 02:47:47 localhost systemd[1]: Started libpod-conmon-9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0.scope. Nov 27 02:47:47 localhost systemd[1]: Started libcrun container. Nov 27 02:47:47 localhost systemd[1]: tmp-crun.7tGCGF.mount: Deactivated successfully. Nov 27 02:47:47 localhost systemd[1]: var-lib-containers-storage-overlay-6ee94bc486d944d874c8a87e1f66717c4ab3f21c16d2b7ac599a9a9e2187bd65-merged.mount: Deactivated successfully. Nov 27 02:47:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.141327276 +0000 UTC m=+0.046518443 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456/merged/var/lib/ceph/osd/ceph-0 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.287529671 +0000 UTC m=+0.192720768 container init 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, description=Red Hat Ceph Storage 7, name=rhceph, GIT_CLEAN=True, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, release=553, RELEASE=main, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, version=7, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public) Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.298103381 +0000 UTC m=+0.203294478 container start 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, io.buildah.version=1.33.12, release=553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, ceph=True, RELEASE=main, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, name=rhceph, GIT_BRANCH=main) Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.298336777 +0000 UTC m=+0.203527864 container attach 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.openshift.expose-services=, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, GIT_CLEAN=True, CEPH_POINT_RELEASE=) Nov 27 02:47:47 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test[31372]: usage: ceph-volume activate [-h] [--osd-id OSD_ID] [--osd-uuid OSD_UUID] Nov 27 02:47:47 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test[31372]: [--no-systemd] [--no-tmpfs] Nov 27 02:47:47 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test[31372]: ceph-volume activate: error: unrecognized arguments: --bad-option Nov 27 02:47:47 localhost systemd[1]: libpod-9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0.scope: Deactivated successfully. Nov 27 02:47:47 localhost podman[31356]: 2025-11-27 07:47:47.521802752 +0000 UTC m=+0.426993909 container died 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, description=Red Hat Ceph Storage 7, name=rhceph, GIT_BRANCH=main, ceph=True, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, RELEASE=main, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, vcs-type=git, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux ) Nov 27 02:47:47 localhost systemd[1]: tmp-crun.qjLfMB.mount: Deactivated successfully. Nov 27 02:47:47 localhost systemd-journald[618]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.1 (250 of 333 items), suggesting rotation. Nov 27 02:47:47 localhost systemd-journald[618]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 02:47:47 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:47:47 localhost systemd[1]: var-lib-containers-storage-overlay-dacac71d4af7d28714161709b8d7bd475b70ae7c69fd8d6b4e57dda3557be456-merged.mount: Deactivated successfully. Nov 27 02:47:47 localhost podman[31377]: 2025-11-27 07:47:47.630987159 +0000 UTC m=+0.095243598 container remove 9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate-test, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, release=553, RELEASE=main, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, distribution-scope=public, maintainer=Guillaume Abrioux ) Nov 27 02:47:47 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:47:47 localhost systemd[1]: libpod-conmon-9956492d603804f600830478f1844566ba64286b3bbb6c29ae4235b556da6aa0.scope: Deactivated successfully. Nov 27 02:47:47 localhost systemd[1]: Reloading. Nov 27 02:47:47 localhost systemd-rc-local-generator[31431]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:47 localhost systemd-sysv-generator[31436]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:48 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:48 localhost systemd[1]: Reloading. Nov 27 02:47:48 localhost systemd-rc-local-generator[31474]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:48 localhost systemd-sysv-generator[31480]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:48 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:48 localhost systemd[1]: Starting Ceph osd.0 for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 02:47:48 localhost podman[31538]: Nov 27 02:47:48 localhost podman[31538]: 2025-11-27 07:47:48.741675881 +0000 UTC m=+0.070043621 container create eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, vcs-type=git, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., version=7, GIT_BRANCH=main, ceph=True, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, architecture=x86_64, CEPH_POINT_RELEASE=, io.openshift.expose-services=, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:48 localhost systemd[1]: tmp-crun.wsg96a.mount: Deactivated successfully. Nov 27 02:47:48 localhost systemd[1]: Started libcrun container. Nov 27 02:47:48 localhost podman[31538]: 2025-11-27 07:47:48.715241694 +0000 UTC m=+0.043609434 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d/merged/var/lib/ceph/osd/ceph-0 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:48 localhost podman[31538]: 2025-11-27 07:47:48.887830104 +0000 UTC m=+0.216197844 container init eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, name=rhceph, description=Red Hat Ceph Storage 7, RELEASE=main, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, architecture=x86_64, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, com.redhat.component=rhceph-container) Nov 27 02:47:48 localhost podman[31538]: 2025-11-27 07:47:48.898063207 +0000 UTC m=+0.226430957 container start eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, io.openshift.tags=rhceph ceph, architecture=x86_64, description=Red Hat Ceph Storage 7, vcs-type=git, maintainer=Guillaume Abrioux , version=7, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, release=553, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, distribution-scope=public, GIT_CLEAN=True, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, RELEASE=main) Nov 27 02:47:48 localhost podman[31538]: 2025-11-27 07:47:48.898354464 +0000 UTC m=+0.226722214 container attach eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, RELEASE=main, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, distribution-scope=public, ceph=True, version=7, name=rhceph, build-date=2025-09-24T08:57:55, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , release=553, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, io.buildah.version=1.33.12, vendor=Red Hat, Inc., vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/ceph-bluestore-tool prime-osd-dir --path /var/lib/ceph/osd/ceph-0 --no-mon-config --dev /dev/mapper/ceph_vg0-ceph_lv0 Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/ceph-bluestore-tool prime-osd-dir --path /var/lib/ceph/osd/ceph-0 --no-mon-config --dev /dev/mapper/ceph_vg0-ceph_lv0 Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/chown -h ceph:ceph /dev/mapper/ceph_vg0-ceph_lv0 Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/chown -h ceph:ceph /dev/mapper/ceph_vg0-ceph_lv0 Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-0 Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-0 Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/ln -s /dev/mapper/ceph_vg0-ceph_lv0 /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/ln -s /dev/mapper/ceph_vg0-ceph_lv0 /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:49 localhost bash[31538]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Nov 27 02:47:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate[31553]: --> ceph-volume raw activate successful for osd ID: 0 Nov 27 02:47:49 localhost bash[31538]: --> ceph-volume raw activate successful for osd ID: 0 Nov 27 02:47:49 localhost systemd[1]: libpod-eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0.scope: Deactivated successfully. Nov 27 02:47:49 localhost podman[31538]: 2025-11-27 07:47:49.623239891 +0000 UTC m=+0.951607631 container died eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, ceph=True, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, io.openshift.expose-services=, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, com.redhat.component=rhceph-container, GIT_CLEAN=True, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:49 localhost podman[31683]: 2025-11-27 07:47:49.714022433 +0000 UTC m=+0.080382316 container remove eda28300b1e3961ce716596e2ccd44f829f1fa96e6a44cdd8dc39dfc8e431fa0 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0-activate, build-date=2025-09-24T08:57:55, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, version=7, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, name=rhceph) Nov 27 02:47:49 localhost systemd[1]: tmp-crun.uRwrck.mount: Deactivated successfully. Nov 27 02:47:49 localhost systemd[1]: var-lib-containers-storage-overlay-01081ec266c87b6d62cd4273711a30026b9a27976c3afe0bd82146e6b94f173d-merged.mount: Deactivated successfully. Nov 27 02:47:50 localhost podman[31742]: Nov 27 02:47:50 localhost podman[31742]: 2025-11-27 07:47:50.021345326 +0000 UTC m=+0.066924467 container create 0c1ed7e21caaa67486f5c2494cce8aafa6907a956a87771607f8fb4297e374cc (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0, distribution-scope=public, ceph=True, GIT_CLEAN=True, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhceph ceph, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, version=7, GIT_BRANCH=main, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=) Nov 27 02:47:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb115c0293c4ef77beb21c9f2b605fa72d44bb357eb742fa6682a7fd854d35db/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb115c0293c4ef77beb21c9f2b605fa72d44bb357eb742fa6682a7fd854d35db/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:50 localhost podman[31742]: 2025-11-27 07:47:49.996866466 +0000 UTC m=+0.042445637 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb115c0293c4ef77beb21c9f2b605fa72d44bb357eb742fa6682a7fd854d35db/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb115c0293c4ef77beb21c9f2b605fa72d44bb357eb742fa6682a7fd854d35db/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb115c0293c4ef77beb21c9f2b605fa72d44bb357eb742fa6682a7fd854d35db/merged/var/lib/ceph/osd/ceph-0 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:50 localhost podman[31742]: 2025-11-27 07:47:50.133409892 +0000 UTC m=+0.178989033 container init 0c1ed7e21caaa67486f5c2494cce8aafa6907a956a87771607f8fb4297e374cc (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0, distribution-scope=public, GIT_CLEAN=True, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-type=git, version=7, architecture=x86_64, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, description=Red Hat Ceph Storage 7, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, GIT_BRANCH=main, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:50 localhost systemd[1]: tmp-crun.cEWB0k.mount: Deactivated successfully. Nov 27 02:47:50 localhost podman[31742]: 2025-11-27 07:47:50.143596152 +0000 UTC m=+0.189175293 container start 0c1ed7e21caaa67486f5c2494cce8aafa6907a956a87771607f8fb4297e374cc (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0, vcs-type=git, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, maintainer=Guillaume Abrioux , version=7, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, architecture=x86_64, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, name=rhceph) Nov 27 02:47:50 localhost bash[31742]: 0c1ed7e21caaa67486f5c2494cce8aafa6907a956a87771607f8fb4297e374cc Nov 27 02:47:50 localhost systemd[1]: Started Ceph osd.0 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 02:47:50 localhost ceph-osd[31760]: set uid:gid to 167:167 (ceph:ceph) Nov 27 02:47:50 localhost ceph-osd[31760]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-osd, pid 2 Nov 27 02:47:50 localhost ceph-osd[31760]: pidfile_write: ignore empty --pid-file Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:50 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:50 localhost ceph-osd[31760]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-0/block size 7.0 GiB Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) close Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) close Nov 27 02:47:50 localhost ceph-osd[31760]: starting osd.0 osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal Nov 27 02:47:50 localhost ceph-osd[31760]: load: jerasure load: lrc Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:50 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:50 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) close Nov 27 02:47:50 localhost podman[31854]: Nov 27 02:47:50 localhost podman[31854]: 2025-11-27 07:47:50.933246545 +0000 UTC m=+0.063255869 container create 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, io.openshift.expose-services=, maintainer=Guillaume Abrioux , RELEASE=main, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, architecture=x86_64, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, vendor=Red Hat, Inc., io.buildah.version=1.33.12, CEPH_POINT_RELEASE=) Nov 27 02:47:50 localhost systemd[1]: Started libpod-conmon-73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba.scope. Nov 27 02:47:50 localhost systemd[1]: Started libcrun container. Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) close Nov 27 02:47:51 localhost podman[31854]: 2025-11-27 07:47:50.905939848 +0000 UTC m=+0.035949172 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:51 localhost podman[31854]: 2025-11-27 07:47:51.004628037 +0000 UTC m=+0.134637371 container init 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, RELEASE=main, vcs-type=git, io.openshift.tags=rhceph ceph, release=553, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, io.openshift.expose-services=, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:51 localhost podman[31854]: 2025-11-27 07:47:51.016959789 +0000 UTC m=+0.146969113 container start 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, architecture=x86_64, name=rhceph, version=7, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7) Nov 27 02:47:51 localhost podman[31854]: 2025-11-27 07:47:51.017274177 +0000 UTC m=+0.147283501 container attach 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, distribution-scope=public, ceph=True, GIT_BRANCH=main, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, name=rhceph, version=7, release=553, io.buildah.version=1.33.12, architecture=x86_64, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, vendor=Red Hat, Inc.) Nov 27 02:47:51 localhost laughing_ellis[31870]: 167 167 Nov 27 02:47:51 localhost systemd[1]: libpod-73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba.scope: Deactivated successfully. Nov 27 02:47:51 localhost podman[31854]: 2025-11-27 07:47:51.019875038 +0000 UTC m=+0.149884382 container died 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , GIT_BRANCH=main, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, name=rhceph, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vendor=Red Hat, Inc., release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55) Nov 27 02:47:51 localhost podman[31880]: 2025-11-27 07:47:51.110425074 +0000 UTC m=+0.081883971 container remove 73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=laughing_ellis, ceph=True, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, vcs-type=git, RELEASE=main, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:51 localhost systemd[1]: libpod-conmon-73253e07a7655adec56c6d62e48e290efec08a9a2ed4ed5bcd6d32a1b7edc6ba.scope: Deactivated successfully. Nov 27 02:47:51 localhost ceph-osd[31760]: mClockScheduler: set_osd_capacity_params_from_config: osd_bandwidth_cost_per_io: 499321.90 bytes/io, osd_bandwidth_capacity_per_shard 157286400.00 bytes/second Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0:0.OSDShard using op scheduler mclock_scheduler, cutoff=196 Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f682e00 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-0/block size 7.0 GiB Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs mount Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs _init_alloc shared, id 1, capacity 0x1bfc00000, block size 0x10000 Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs mount shared_bdev_used = 0 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _prepare_db_environment set db_paths to db,7136398540 db.slow,7136398540 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: RocksDB version: 7.9.2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Git sha 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DB SUMMARY Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DB Session ID: VFEN9H8ZYWUSKOB3D7NP Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: CURRENT file: CURRENT Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: IDENTITY file: IDENTITY Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: MANIFEST file: MANIFEST-000032 size: 1007 Bytes Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: SST files in db dir, Total Num: 1, files: 000030.sst Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: SST files in db.slow dir, Total Num: 0, files: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Write Ahead Log file in db.wal: 000031.log size: 5093 ; Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.error_if_exists: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.create_if_missing: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.env: 0x56526f916cb0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.fs: LegacyFileSystem Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.info_log: 0x565270604380 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.statistics: (nil) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_fsync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_log_file_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_fallocate: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_direct_reads: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.create_missing_column_families: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_log_dir: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_dir: db.wal Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.advise_random_on_open: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_manager: 0x56526f66c140 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.rate_limiter: (nil) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.unordered_write: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.row_cache: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.two_write_queues: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.manual_wal_flush: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_compression: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.atomic_flush: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.log_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_host_id: __hostname__ Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_jobs: 4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_compactions: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_subcompactions: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.writable_file_max_buffer_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_total_wal_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_open_files: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_readahead_size: 2097152 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_flushes: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Compression algorithms supported: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZSTD supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kXpressCompression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZlibCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_readonly.cc:25] Opening the db in read only mode Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: db/MANIFEST-000032 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 0, name: default) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: .T:int64_array.b:bitwise_xor Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 1, name: m-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 2, name: m-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 3, name: m-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 4, name: p-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 5, name: p-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 6, name: p-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604540)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 7, name: O-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604760)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 8, name: O-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604760)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 9, name: O-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x565270604760)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 10, name: L) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 11, name: P) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:db/MANIFEST-000032 succeeded,manifest_file_number is 32, next_file_number is 34, last_sequence is 12, log_number is 5,prev_log_number is 0,max_column_family is 11,min_log_number_to_keep is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-0] (ID 1), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-1] (ID 2), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-2] (ID 3), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-0] (ID 4), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-1] (ID 5), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-2] (ID 6), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-0] (ID 7), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-1] (ID 8), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-2] (ID 9), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [L] (ID 10), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [P] (ID 11), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ac6665ef-7364-4832-b285-e8c95d6c85dc Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671305600, "job": 1, "event": "recovery_started", "wal_files": [31]} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #31 mode 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671305897, "job": 1, "event": "recovery_finished"} Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_db opened rocksdb path db options compression=kLZ4Compression,max_write_buffer_number=64,min_write_buffer_number_to_merge=6,compaction_style=kCompactionStyleLevel,write_buffer_size=16777216,max_background_jobs=4,level0_file_num_compaction_trigger=8,max_bytes_for_level_base=1073741824,max_bytes_for_level_multiplier=8,compaction_readahead_size=2MB,max_total_wal_size=1073741824,writable_file_max_buffer_size=0 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_super_meta old nid_max 1025 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_super_meta old blobid_max 10240 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_super_meta ondisk_format 4 compat_ondisk_format 3 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_super_meta min_alloc_size 0x1000 Nov 27 02:47:51 localhost ceph-osd[31760]: freelist init Nov 27 02:47:51 localhost ceph-osd[31760]: freelist _read_cfg Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _init_alloc loaded 7.0 GiB in 2 extents, allocator type hybrid, capacity 0x1bfc00000, block size 0x1000, free 0x1bfbfd000, fragmentation 5.5e-07 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs umount Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) close Nov 27 02:47:51 localhost podman[32103]: Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.437363922 +0000 UTC m=+0.072057559 container create 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, GIT_BRANCH=main, architecture=x86_64, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, ceph=True, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, release=553, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , GIT_CLEAN=True, description=Red Hat Ceph Storage 7, version=7, RELEASE=main, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:51 localhost systemd[1]: Started libpod-conmon-833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7.scope. Nov 27 02:47:51 localhost systemd[1]: Started libcrun container. Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.408719383 +0000 UTC m=+0.043413060 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:51 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0cfa316b0524b6e193f2b192897e6a8531e659e42c19dbae5bfef4b32144092b/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:51 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0cfa316b0524b6e193f2b192897e6a8531e659e42c19dbae5bfef4b32144092b/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:51 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0cfa316b0524b6e193f2b192897e6a8531e659e42c19dbae5bfef4b32144092b/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:51 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0cfa316b0524b6e193f2b192897e6a8531e659e42c19dbae5bfef4b32144092b/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:51 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0cfa316b0524b6e193f2b192897e6a8531e659e42c19dbae5bfef4b32144092b/merged/var/lib/ceph/osd/ceph-3 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.566550953 +0000 UTC m=+0.201244600 container init 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, version=7, CEPH_POINT_RELEASE=, ceph=True, architecture=x86_64, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , release=553, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.openshift.expose-services=) Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.576343276 +0000 UTC m=+0.211036933 container start 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, GIT_CLEAN=True, release=553, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, GIT_BRANCH=main, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7) Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.576663133 +0000 UTC m=+0.211356840 container attach 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., version=7, GIT_BRANCH=main, distribution-scope=public, ceph=True, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , RELEASE=main, CEPH_POINT_RELEASE=, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, GIT_CLEAN=True, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open path /var/lib/ceph/osd/ceph-0/block Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-0/block failed: (22) Invalid argument Nov 27 02:47:51 localhost ceph-osd[31760]: bdev(0x56526f683180 /var/lib/ceph/osd/ceph-0/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-0/block size 7.0 GiB Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs mount Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs _init_alloc shared, id 1, capacity 0x1bfc00000, block size 0x10000 Nov 27 02:47:51 localhost ceph-osd[31760]: bluefs mount shared_bdev_used = 4718592 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _prepare_db_environment set db_paths to db,7136398540 db.slow,7136398540 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: RocksDB version: 7.9.2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Git sha 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DB SUMMARY Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DB Session ID: VFEN9H8ZYWUSKOB3D7NO Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: CURRENT file: CURRENT Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: IDENTITY file: IDENTITY Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: MANIFEST file: MANIFEST-000032 size: 1007 Bytes Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: SST files in db dir, Total Num: 1, files: 000030.sst Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: SST files in db.slow dir, Total Num: 0, files: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Write Ahead Log file in db.wal: 000031.log size: 5093 ; Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.error_if_exists: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.create_if_missing: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.env: 0x56526f917ea0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.fs: LegacyFileSystem Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.info_log: 0x565270605480 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.statistics: (nil) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_fsync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_log_file_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_fallocate: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_direct_reads: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.create_missing_column_families: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_log_dir: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_dir: db.wal Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.advise_random_on_open: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_manager: 0x56526f66d540 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.rate_limiter: (nil) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.unordered_write: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.row_cache: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.two_write_queues: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.manual_wal_flush: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_compression: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.atomic_flush: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.log_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.db_host_id: __hostname__ Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_jobs: 4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_compactions: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_subcompactions: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.writable_file_max_buffer_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_total_wal_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_open_files: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_readahead_size: 2097152 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_background_flushes: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Compression algorithms supported: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZSTD supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kXpressCompression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kZlibCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: db/MANIFEST-000032 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 0, name: default) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: .T:int64_array.b:bitwise_xor Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 1, name: m-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 2, name: m-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 3, name: m-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 4, name: p-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 5, name: p-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 6, name: p-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8a00)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65b610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 7, name: O-0) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-0]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8bc0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 8, name: O-1) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-1]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8bc0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 9, name: O-2) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-2]: Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.merge_operator: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5652706a8bc0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56526f65a2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression: LZ4 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.num_levels: 7 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 10, name: L) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 11, name: P) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:db/MANIFEST-000032 succeeded,manifest_file_number is 32, next_file_number is 34, last_sequence is 12, log_number is 5,prev_log_number is 0,max_column_family is 11,min_log_number_to_keep is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-0] (ID 1), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-1] (ID 2), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [m-2] (ID 3), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-0] (ID 4), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-1] (ID 5), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [p-2] (ID 6), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-0] (ID 7), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-1] (ID 8), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [O-2] (ID 9), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [L] (ID 10), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5581] Column family [P] (ID 11), log number is 5 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ac6665ef-7364-4832-b285-e8c95d6c85dc Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671630370, "job": 1, "event": "recovery_started", "wal_files": [31]} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #31 mode 2 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671638227, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 35, "file_size": 1261, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 13, "largest_seqno": 21, "table_properties": {"data_size": 128, "index_size": 27, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 87, "raw_average_key_size": 17, "raw_value_size": 82, "raw_average_value_size": 16, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 2, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": ".T:int64_array.b:bitwise_xor", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229671, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ac6665ef-7364-4832-b285-e8c95d6c85dc", "db_session_id": "VFEN9H8ZYWUSKOB3D7NO", "orig_file_number": 35, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671643153, "cf_name": "p-0", "job": 1, "event": "table_file_creation", "file_number": 36, "file_size": 1609, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 14, "largest_seqno": 15, "table_properties": {"data_size": 468, "index_size": 39, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 72, "raw_average_key_size": 36, "raw_value_size": 567, "raw_average_value_size": 283, "num_data_blocks": 1, "num_entries": 2, "num_filter_entries": 2, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "p-0", "column_family_id": 4, "comparator": "leveldb.BytewiseComparator", "merge_operator": "nullptr", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229671, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ac6665ef-7364-4832-b285-e8c95d6c85dc", "db_session_id": "VFEN9H8ZYWUSKOB3D7NO", "orig_file_number": 36, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671646972, "cf_name": "O-2", "job": 1, "event": "table_file_creation", "file_number": 37, "file_size": 1290, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 16, "largest_seqno": 16, "table_properties": {"data_size": 121, "index_size": 64, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 55, "raw_average_key_size": 55, "raw_value_size": 50, "raw_average_value_size": 50, "num_data_blocks": 1, "num_entries": 1, "num_filter_entries": 1, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "O-2", "column_family_id": 9, "comparator": "leveldb.BytewiseComparator", "merge_operator": "nullptr", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229671, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ac6665ef-7364-4832-b285-e8c95d6c85dc", "db_session_id": "VFEN9H8ZYWUSKOB3D7NO", "orig_file_number": 37, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:1432] Failed to truncate log #31: IO error: No such file or directory: While open a file for appending: db.wal/000031.log: No such file or directory Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229671650439, "job": 1, "event": "recovery_finished"} Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/version_set.cc:5047] Creating manifest 40 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5652706b4380 Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: DB pointer 0x56527055ba00 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _open_db opened rocksdb path db options compression=kLZ4Compression,max_write_buffer_number=64,min_write_buffer_number_to_merge=6,compaction_style=kCompactionStyleLevel,write_buffer_size=16777216,max_background_jobs=4,level0_file_num_compaction_trigger=8,max_bytes_for_level_base=1073741824,max_bytes_for_level_multiplier=8,compaction_readahead_size=2MB,max_total_wal_size=1073741824,writable_file_max_buffer_size=0 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _upgrade_super from 4, latest 4 Nov 27 02:47:51 localhost ceph-osd[31760]: bluestore(/var/lib/ceph/osd/ceph-0) _upgrade_super done Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 02:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 0.1 total, 0.1 interval#012Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s#012Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.1 total, 0.1 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 460.80 MB usage: 1.39 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 8 last_secs: 4.1e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(2,0.72 KB,0.000152323%) FilterBlock(3,0.33 KB,6.95388e-05%) IndexBlock(3,0.34 KB,7.28501e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.1 total, 0.1 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 460.80 MB usage: 1.39 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 8 last_secs: 4.1e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(2,0.72 KB,0.000152323%) FilterBlock(3,0.33 KB,6.95388e-05%) IndexBlock(3,0.34 KB,7.28501e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.1 total, 0.1 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012 Nov 27 02:47:51 localhost ceph-osd[31760]: /builddir/build/BUILD/ceph-18.2.1/src/cls/cephfs/cls_cephfs.cc:201: loading cephfs Nov 27 02:47:51 localhost ceph-osd[31760]: /builddir/build/BUILD/ceph-18.2.1/src/cls/hello/cls_hello.cc:316: loading cls_hello Nov 27 02:47:51 localhost ceph-osd[31760]: _get_class not permitted to load lua Nov 27 02:47:51 localhost ceph-osd[31760]: _get_class not permitted to load sdk Nov 27 02:47:51 localhost ceph-osd[31760]: _get_class not permitted to load test_remote_reads Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 crush map has features 288232575208783872, adjusting msgr requires for clients Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 crush map has features 288232575208783872 was 8705, adjusting msgr requires for mons Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 crush map has features 288232575208783872, adjusting msgr requires for osds Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 check_osdmap_features enabling on-disk ERASURE CODES compat feature Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 load_pgs Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 load_pgs opened 0 pgs Nov 27 02:47:51 localhost ceph-osd[31760]: osd.0 0 log_to_monitors true Nov 27 02:47:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0[31756]: 2025-11-27T07:47:51.687+0000 7f8f20060a80 -1 osd.0 0 log_to_monitors true Nov 27 02:47:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test[32118]: usage: ceph-volume activate [-h] [--osd-id OSD_ID] [--osd-uuid OSD_UUID] Nov 27 02:47:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test[32118]: [--no-systemd] [--no-tmpfs] Nov 27 02:47:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test[32118]: ceph-volume activate: error: unrecognized arguments: --bad-option Nov 27 02:47:51 localhost systemd[1]: libpod-833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7.scope: Deactivated successfully. Nov 27 02:47:51 localhost podman[32103]: 2025-11-27 07:47:51.804983053 +0000 UTC m=+0.439676660 container died 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, release=553, RELEASE=main, build-date=2025-09-24T08:57:55, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, version=7, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.buildah.version=1.33.12, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64) Nov 27 02:47:51 localhost podman[32339]: 2025-11-27 07:47:51.872741829 +0000 UTC m=+0.059240785 container remove 833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate-test, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, vcs-type=git, io.buildah.version=1.33.12, GIT_BRANCH=main, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, io.openshift.tags=rhceph ceph, version=7) Nov 27 02:47:51 localhost systemd[1]: libpod-conmon-833f2d16028a0bde77edc460e3fc9bccf456c92ca2ad16394e8c30b9d74e36f7.scope: Deactivated successfully. Nov 27 02:47:51 localhost systemd[1]: var-lib-containers-storage-overlay-b55a7dd56b79c5bfc6654d733a97406c1a2c05d71fb2f8747cd34ed2d0266ef2-merged.mount: Deactivated successfully. Nov 27 02:47:52 localhost systemd[1]: Reloading. Nov 27 02:47:52 localhost systemd-rc-local-generator[32395]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:52 localhost systemd-sysv-generator[32399]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:52 localhost systemd[1]: Reloading. Nov 27 02:47:52 localhost systemd-sysv-generator[32438]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:47:52 localhost systemd-rc-local-generator[32434]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:47:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:47:52 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : purged_snaps scrub starts Nov 27 02:47:52 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : purged_snaps scrub ok Nov 27 02:47:52 localhost systemd[1]: Starting Ceph osd.3 for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 done with init, starting boot process Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 start_boot Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 maybe_override_options_for_qos osd_max_backfills set to 1 Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 maybe_override_options_for_qos osd_recovery_max_active set to 0 Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 maybe_override_options_for_qos osd_recovery_max_active_hdd set to 3 Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 maybe_override_options_for_qos osd_recovery_max_active_ssd set to 10 Nov 27 02:47:52 localhost ceph-osd[31760]: osd.0 0 bench count 12288000 bsize 4 KiB Nov 27 02:47:53 localhost podman[32498]: Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.053855309 +0000 UTC m=+0.078903321 container create 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, description=Red Hat Ceph Storage 7, name=rhceph, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, ceph=True, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, maintainer=Guillaume Abrioux , version=7, vendor=Red Hat, Inc., vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, RELEASE=main) Nov 27 02:47:53 localhost systemd[1]: Started libcrun container. Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.014431455 +0000 UTC m=+0.039479477 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616/merged/var/lib/ceph/osd/ceph-3 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.172979772 +0000 UTC m=+0.198027754 container init 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, name=rhceph, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, vcs-type=git, release=553, maintainer=Guillaume Abrioux , version=7, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.183828628 +0000 UTC m=+0.208876640 container start 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, distribution-scope=public, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, architecture=x86_64, io.buildah.version=1.33.12, GIT_CLEAN=True, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, io.openshift.expose-services=, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7) Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.184769551 +0000 UTC m=+0.209817543 container attach 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, distribution-scope=public, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, maintainer=Guillaume Abrioux , version=7, ceph=True, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, io.openshift.expose-services=) Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/ceph-bluestore-tool prime-osd-dir --path /var/lib/ceph/osd/ceph-3 --no-mon-config --dev /dev/mapper/ceph_vg1-ceph_lv1 Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/ceph-bluestore-tool prime-osd-dir --path /var/lib/ceph/osd/ceph-3 --no-mon-config --dev /dev/mapper/ceph_vg1-ceph_lv1 Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/chown -h ceph:ceph /dev/mapper/ceph_vg1-ceph_lv1 Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/chown -h ceph:ceph /dev/mapper/ceph_vg1-ceph_lv1 Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/ln -s /dev/mapper/ceph_vg1-ceph_lv1 /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/ln -s /dev/mapper/ceph_vg1-ceph_lv1 /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:53 localhost bash[32498]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 Nov 27 02:47:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate[32513]: --> ceph-volume raw activate successful for osd ID: 3 Nov 27 02:47:53 localhost bash[32498]: --> ceph-volume raw activate successful for osd ID: 3 Nov 27 02:47:53 localhost systemd[1]: libpod-8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d.scope: Deactivated successfully. Nov 27 02:47:53 localhost podman[32498]: 2025-11-27 07:47:53.938626556 +0000 UTC m=+0.963674578 container died 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, maintainer=Guillaume Abrioux , ceph=True, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, name=rhceph, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., architecture=x86_64, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main) Nov 27 02:47:54 localhost systemd[1]: var-lib-containers-storage-overlay-12fc471df37fa146cb5e59c4b76d68f272f2f96a06dbf111271c4e8f90f3f616-merged.mount: Deactivated successfully. Nov 27 02:47:54 localhost podman[32644]: 2025-11-27 07:47:54.07762941 +0000 UTC m=+0.127581455 container remove 8c871508ec81374d596b4710483b5573eda073b0fb710fce2e5388b97c698c5d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3-activate, architecture=x86_64, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, version=7, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, name=rhceph, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, GIT_CLEAN=True, release=553, maintainer=Guillaume Abrioux , vcs-type=git, GIT_BRANCH=main, distribution-scope=public, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, RELEASE=main) Nov 27 02:47:54 localhost podman[32704]: Nov 27 02:47:54 localhost podman[32704]: 2025-11-27 07:47:54.387642336 +0000 UTC m=+0.076817641 container create 8d4bee6b81dd72fba3f00434691f819e18a4f9c303cbf62489ea1536e141d78c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, GIT_BRANCH=main, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, name=rhceph, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, architecture=x86_64, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553) Nov 27 02:47:54 localhost podman[32704]: 2025-11-27 07:47:54.356704063 +0000 UTC m=+0.045879388 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1877aafc5b33de8e06082ae5a236a4b8333af5a63605d18173e7bc56274e9f0e/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1877aafc5b33de8e06082ae5a236a4b8333af5a63605d18173e7bc56274e9f0e/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1877aafc5b33de8e06082ae5a236a4b8333af5a63605d18173e7bc56274e9f0e/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1877aafc5b33de8e06082ae5a236a4b8333af5a63605d18173e7bc56274e9f0e/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1877aafc5b33de8e06082ae5a236a4b8333af5a63605d18173e7bc56274e9f0e/merged/var/lib/ceph/osd/ceph-3 supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:54 localhost podman[32704]: 2025-11-27 07:47:54.549513942 +0000 UTC m=+0.238689247 container init 8d4bee6b81dd72fba3f00434691f819e18a4f9c303cbf62489ea1536e141d78c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3, build-date=2025-09-24T08:57:55, distribution-scope=public, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, io.buildah.version=1.33.12, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, release=553, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, RELEASE=main, vendor=Red Hat, Inc.) Nov 27 02:47:54 localhost podman[32704]: 2025-11-27 07:47:54.586944169 +0000 UTC m=+0.276119474 container start 8d4bee6b81dd72fba3f00434691f819e18a4f9c303cbf62489ea1536e141d78c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3, version=7, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, distribution-scope=public, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, RELEASE=main, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, GIT_BRANCH=main, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, ceph=True, vcs-type=git) Nov 27 02:47:54 localhost bash[32704]: 8d4bee6b81dd72fba3f00434691f819e18a4f9c303cbf62489ea1536e141d78c Nov 27 02:47:54 localhost systemd[1]: Started Ceph osd.3 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 02:47:54 localhost ceph-osd[32722]: set uid:gid to 167:167 (ceph:ceph) Nov 27 02:47:54 localhost ceph-osd[32722]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-osd, pid 2 Nov 27 02:47:54 localhost ceph-osd[32722]: pidfile_write: ignore empty --pid-file Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:54 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:54 localhost ceph-osd[32722]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-3/block size 7.0 GiB Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) close Nov 27 02:47:54 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) close Nov 27 02:47:55 localhost ceph-osd[32722]: starting osd.3 osd_data /var/lib/ceph/osd/ceph-3 /var/lib/ceph/osd/ceph-3/journal Nov 27 02:47:55 localhost ceph-osd[32722]: load: jerasure load: lrc Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) close Nov 27 02:47:55 localhost podman[32809]: Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) close Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.426003733 +0000 UTC m=+0.078050151 container create 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, name=rhceph, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, vendor=Red Hat, Inc., ceph=True, io.openshift.tags=rhceph ceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_BRANCH=main, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux ) Nov 27 02:47:55 localhost systemd[1]: Started libpod-conmon-3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324.scope. Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.390774068 +0000 UTC m=+0.042820506 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:55 localhost systemd[1]: Started libcrun container. Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.522353106 +0000 UTC m=+0.174399524 container init 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, ceph=True, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, io.openshift.tags=rhceph ceph, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, version=7, io.openshift.expose-services=, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, vendor=Red Hat, Inc., GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, vcs-type=git, name=rhceph, description=Red Hat Ceph Storage 7) Nov 27 02:47:55 localhost condescending_moore[32828]: 167 167 Nov 27 02:47:55 localhost systemd[1]: libpod-3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324.scope: Deactivated successfully. Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.558980384 +0000 UTC m=+0.211026832 container start 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, version=7, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-type=git, architecture=x86_64, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.560183103 +0000 UTC m=+0.212229571 container attach 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, architecture=x86_64, release=553, RELEASE=main, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, distribution-scope=public, GIT_BRANCH=main, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container) Nov 27 02:47:55 localhost podman[32809]: 2025-11-27 07:47:55.566571644 +0000 UTC m=+0.218618082 container died 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_CLEAN=True, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, vendor=Red Hat, Inc., GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, release=553, io.buildah.version=1.33.12, ceph=True, maintainer=Guillaume Abrioux ) Nov 27 02:47:55 localhost ceph-osd[32722]: mClockScheduler: set_osd_capacity_params_from_config: osd_bandwidth_cost_per_io: 499321.90 bytes/io, osd_bandwidth_capacity_per_shard 157286400.00 bytes/second Nov 27 02:47:55 localhost ceph-osd[32722]: osd.3:0.OSDShard using op scheduler mclock_scheduler, cutoff=196 Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b4e00 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06 Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-3/block size 7.0 GiB Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs mount Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs _init_alloc shared, id 1, capacity 0x1bfc00000, block size 0x10000 Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs mount shared_bdev_used = 0 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _prepare_db_environment set db_paths to db,7136398540 db.slow,7136398540 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: RocksDB version: 7.9.2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Git sha 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DB SUMMARY Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DB Session ID: EWHJQ1K89QT3T8EADWXL Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: CURRENT file: CURRENT Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: IDENTITY file: IDENTITY Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: MANIFEST file: MANIFEST-000032 size: 1007 Bytes Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: SST files in db dir, Total Num: 1, files: 000030.sst Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: SST files in db.slow dir, Total Num: 0, files: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Write Ahead Log file in db.wal: 000031.log size: 5093 ; Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.error_if_exists: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.create_if_missing: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.env: 0x5568f6448cb0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.fs: LegacyFileSystem Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.info_log: 0x5568f7132780 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.statistics: (nil) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_fsync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_log_file_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_fallocate: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_direct_reads: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.create_missing_column_families: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_log_dir: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_dir: db.wal Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.advise_random_on_open: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_manager: 0x5568f619e140 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.rate_limiter: (nil) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.unordered_write: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.row_cache: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.two_write_queues: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.manual_wal_flush: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_compression: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.atomic_flush: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.log_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_host_id: __hostname__ Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_jobs: 4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_compactions: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_subcompactions: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.writable_file_max_buffer_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_total_wal_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_open_files: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_readahead_size: 2097152 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_flushes: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Compression algorithms supported: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZSTD supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kXpressCompression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZlibCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_readonly.cc:25] Opening the db in read only mode Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: db/MANIFEST-000032 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 0, name: default) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: .T:int64_array.b:bitwise_xor Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost podman[32833]: 2025-11-27 07:47:55.728189824 +0000 UTC m=+0.177021455 container remove 3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_moore, io.openshift.expose-services=, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., release=553, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, name=rhceph, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, ceph=True, GIT_CLEAN=True) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost systemd[1]: libpod-conmon-3a599d3c7cf92a3239a9f5d0273f9cedd4e82eb099757a944435c92d34f78324.scope: Deactivated successfully. Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 1, name: m-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 2, name: m-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 3, name: m-2) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-2]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 4, name: p-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 5, name: p-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 6, name: p-2) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-2]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132940)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c850#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 7, name: O-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132b60)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 8, name: O-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132b60)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 9, name: O-2) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-2]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f7132b60)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 10, name: L) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 11, name: P) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:db/MANIFEST-000032 succeeded,manifest_file_number is 32, next_file_number is 34, last_sequence is 12, log_number is 5,prev_log_number is 0,max_column_family is 11,min_log_number_to_keep is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-0] (ID 1), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-1] (ID 2), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-2] (ID 3), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-0] (ID 4), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-1] (ID 5), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-2] (ID 6), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-0] (ID 7), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-1] (ID 8), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-2] (ID 9), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [L] (ID 10), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [P] (ID 11), log number is 5 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 91651a6c-8384-475b-b1ef-ba200d331a62 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229675735449, "job": 1, "event": "recovery_started", "wal_files": [31]} Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #31 mode 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229675735752, "job": 1, "event": "recovery_finished"} Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_db opened rocksdb path db options compression=kLZ4Compression,max_write_buffer_number=64,min_write_buffer_number_to_merge=6,compaction_style=kCompactionStyleLevel,write_buffer_size=16777216,max_background_jobs=4,level0_file_num_compaction_trigger=8,max_bytes_for_level_base=1073741824,max_bytes_for_level_multiplier=8,compaction_readahead_size=2MB,max_total_wal_size=1073741824,writable_file_max_buffer_size=0 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_super_meta old nid_max 1025 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_super_meta old blobid_max 10240 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_super_meta ondisk_format 4 compat_ondisk_format 3 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_super_meta min_alloc_size 0x1000 Nov 27 02:47:55 localhost ceph-osd[32722]: freelist init Nov 27 02:47:55 localhost ceph-osd[32722]: freelist _read_cfg Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _init_alloc loaded 7.0 GiB in 2 extents, allocator type hybrid, capacity 0x1bfc00000, block size 0x1000, free 0x1bfbfd000, fragmentation 5.5e-07 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs umount Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) close Nov 27 02:47:55 localhost podman[33048]: Nov 27 02:47:55 localhost podman[33048]: 2025-11-27 07:47:55.936808448 +0000 UTC m=+0.086049561 container create 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, io.openshift.expose-services=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, distribution-scope=public, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, architecture=x86_64, release=553, ceph=True, CEPH_POINT_RELEASE=, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph) Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open path /var/lib/ceph/osd/ceph-3/block Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) ioctl(F_SET_FILE_RW_HINT) on /var/lib/ceph/osd/ceph-3/block failed: (22) Invalid argument Nov 27 02:47:55 localhost ceph-osd[32722]: bdev(0x5568f61b5180 /var/lib/ceph/osd/ceph-3/block) open size 7511998464 (0x1bfc00000, 7.0 GiB) block_size 4096 (4 KiB) rotational device, discard supported Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs add_block_device bdev 1 path /var/lib/ceph/osd/ceph-3/block size 7.0 GiB Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs mount Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs _init_alloc shared, id 1, capacity 0x1bfc00000, block size 0x10000 Nov 27 02:47:55 localhost ceph-osd[32722]: bluefs mount shared_bdev_used = 4718592 Nov 27 02:47:55 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _prepare_db_environment set db_paths to db,7136398540 db.slow,7136398540 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: RocksDB version: 7.9.2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Git sha 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DB SUMMARY Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DB Session ID: EWHJQ1K89QT3T8EADWXK Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: CURRENT file: CURRENT Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: IDENTITY file: IDENTITY Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: MANIFEST file: MANIFEST-000032 size: 1007 Bytes Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: SST files in db dir, Total Num: 1, files: 000030.sst Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: SST files in db.slow dir, Total Num: 0, files: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Write Ahead Log file in db.wal: 000031.log size: 5093 ; Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.error_if_exists: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.create_if_missing: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.env: 0x5568f61f2460 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.fs: LegacyFileSystem Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.info_log: 0x5568f7133640 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.statistics: (nil) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_fsync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_log_file_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_fallocate: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_direct_reads: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.create_missing_column_families: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_log_dir: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_dir: db.wal Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.advise_random_on_open: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_manager: 0x5568f619f5e0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.rate_limiter: (nil) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.unordered_write: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.row_cache: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.two_write_queues: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.manual_wal_flush: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_compression: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.atomic_flush: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.log_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.db_host_id: __hostname__ Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_jobs: 4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_compactions: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_subcompactions: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.writable_file_max_buffer_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_total_wal_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_open_files: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_readahead_size: 2097152 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_background_flushes: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Compression algorithms supported: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZSTD supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kXpressCompression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kZlibCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: db/MANIFEST-000032 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 0, name: default) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: .T:int64_array.b:bitwise_xor Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 1, name: m-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 2, name: m-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 3, name: m-2) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [m-2]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 4, name: p-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 5, name: p-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost systemd[1]: Started libpod-conmon-415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e.scope. Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 6, name: p-2) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [p-2]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71ab3a0)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618c2d0#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 483183820#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 7, name: O-0) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-0]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71aa740)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618d610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 8, name: O-1) Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-1]: Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71aa740)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618d610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:55 localhost podman[33048]: 2025-11-27 07:47:55.894443664 +0000 UTC m=+0.043684777 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:55 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 9, name: O-2) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [O-2]: Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.merge_operator: None Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter: None Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_filter_factory: None Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.sst_partitioner_factory: None Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5568f71aa740)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x5568f618d610#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.write_buffer_size: 16777216 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number: 64 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression: LZ4 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression: Disabled Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.prefix_extractor: nullptr Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.num_levels: 7 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.min_write_buffer_number_to_merge: 6 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.level: 32767 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.enabled: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_file_num_compaction_trigger: 8 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_base: 1073741824 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.level_compaction_dynamic_level_bytes: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier: 8.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.arena_block_size: 1048576 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_support: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.bloom_locality: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.max_successive_merges: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.force_consistency_checks: 1 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.ttl: 2592000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_files: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.min_blob_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_file_size: 268435456 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 10, name: L) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:578] Failed to register data paths of column family (id: 11, name: P) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/column_family.cc:635] #011(skipping printing options) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:db/MANIFEST-000032 succeeded,manifest_file_number is 32, next_file_number is 34, last_sequence is 12, log_number is 5,prev_log_number is 0,max_column_family is 11,min_log_number_to_keep is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-0] (ID 1), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-1] (ID 2), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [m-2] (ID 3), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-0] (ID 4), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-1] (ID 5), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [p-2] (ID 6), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-0] (ID 7), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-1] (ID 8), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [O-2] (ID 9), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [L] (ID 10), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5581] Column family [P] (ID 11), log number is 5 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 91651a6c-8384-475b-b1ef-ba200d331a62 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229675999297, "job": 1, "event": "recovery_started", "wal_files": [31]} Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #31 mode 2 Nov 27 02:47:56 localhost systemd[1]: Started libcrun container. Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229676016374, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 35, "file_size": 1261, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 13, "largest_seqno": 21, "table_properties": {"data_size": 128, "index_size": 27, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 87, "raw_average_key_size": 17, "raw_value_size": 82, "raw_average_value_size": 16, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 2, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": ".T:int64_array.b:bitwise_xor", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229675, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "91651a6c-8384-475b-b1ef-ba200d331a62", "db_session_id": "EWHJQ1K89QT3T8EADWXK", "orig_file_number": 35, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:56 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4c03b0b894c24eeb55f59c6c056222fa2f980ce4307e1b3cf02ea1a26ea8643e/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:56 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4c03b0b894c24eeb55f59c6c056222fa2f980ce4307e1b3cf02ea1a26ea8643e/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229676039919, "cf_name": "p-0", "job": 1, "event": "table_file_creation", "file_number": 36, "file_size": 1607, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 14, "largest_seqno": 15, "table_properties": {"data_size": 466, "index_size": 39, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 72, "raw_average_key_size": 36, "raw_value_size": 567, "raw_average_value_size": 283, "num_data_blocks": 1, "num_entries": 2, "num_filter_entries": 2, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "p-0", "column_family_id": 4, "comparator": "leveldb.BytewiseComparator", "merge_operator": "nullptr", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229676, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "91651a6c-8384-475b-b1ef-ba200d331a62", "db_session_id": "EWHJQ1K89QT3T8EADWXK", "orig_file_number": 36, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229676047744, "cf_name": "O-2", "job": 1, "event": "table_file_creation", "file_number": 37, "file_size": 1290, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 16, "largest_seqno": 16, "table_properties": {"data_size": 121, "index_size": 64, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 55, "raw_average_key_size": 55, "raw_value_size": 50, "raw_average_value_size": 50, "num_data_blocks": 1, "num_entries": 1, "num_filter_entries": 1, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "O-2", "column_family_id": 9, "comparator": "leveldb.BytewiseComparator", "merge_operator": "nullptr", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "LZ4", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764229676, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "91651a6c-8384-475b-b1ef-ba200d331a62", "db_session_id": "EWHJQ1K89QT3T8EADWXK", "orig_file_number": 37, "seqno_to_time_mapping": "N/A"}} Nov 27 02:47:56 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4c03b0b894c24eeb55f59c6c056222fa2f980ce4307e1b3cf02ea1a26ea8643e/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:1432] Failed to truncate log #31: IO error: No such file or directory: While open a file for appending: db.wal/000031.log: No such file or directory Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764229676063680, "job": 1, "event": "recovery_finished"} Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/version_set.cc:5047] Creating manifest 40 Nov 27 02:47:56 localhost podman[33048]: 2025-11-27 07:47:56.064381181 +0000 UTC m=+0.213622264 container init 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, distribution-scope=public, RELEASE=main, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, ceph=True, name=rhceph, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:47:56 localhost podman[33048]: 2025-11-27 07:47:56.084071648 +0000 UTC m=+0.233312721 container start 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, name=rhceph, io.openshift.expose-services=, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, distribution-scope=public, architecture=x86_64, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:47:56 localhost podman[33048]: 2025-11-27 07:47:56.084262872 +0000 UTC m=+0.233503965 container attach 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, io.buildah.version=1.33.12, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_CLEAN=True, vcs-type=git, architecture=x86_64, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, GIT_BRANCH=main, io.openshift.expose-services=, RELEASE=main, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, release=553, ceph=True) Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5568f623c700 Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: DB pointer 0x5568f7089a00 Nov 27 02:47:56 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _open_db opened rocksdb path db options compression=kLZ4Compression,max_write_buffer_number=64,min_write_buffer_number_to_merge=6,compaction_style=kCompactionStyleLevel,write_buffer_size=16777216,max_background_jobs=4,level0_file_num_compaction_trigger=8,max_bytes_for_level_base=1073741824,max_bytes_for_level_multiplier=8,compaction_readahead_size=2MB,max_total_wal_size=1073741824,writable_file_max_buffer_size=0 Nov 27 02:47:56 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _upgrade_super from 4, latest 4 Nov 27 02:47:56 localhost ceph-osd[32722]: bluestore(/var/lib/ceph/osd/ceph-3) _upgrade_super done Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 02:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 0.2 total, 0.2 interval#012Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s#012Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.2 total, 0.2 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 460.80 MB usage: 0.67 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 8 last_secs: 3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): FilterBlock(3,0.33 KB,6.95388e-05%) IndexBlock(3,0.34 KB,7.28501e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.2 total, 0.2 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 460.80 MB usage: 0.67 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 8 last_secs: 3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): FilterBlock(3,0.33 KB,6.95388e-05%) IndexBlock(3,0.34 KB,7.28501e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.2 total, 0.2 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 460.80 MB usage: 0 Nov 27 02:47:56 localhost ceph-osd[32722]: /builddir/build/BUILD/ceph-18.2.1/src/cls/cephfs/cls_cephfs.cc:201: loading cephfs Nov 27 02:47:56 localhost ceph-osd[32722]: /builddir/build/BUILD/ceph-18.2.1/src/cls/hello/cls_hello.cc:316: loading cls_hello Nov 27 02:47:56 localhost ceph-osd[32722]: _get_class not permitted to load lua Nov 27 02:47:56 localhost ceph-osd[32722]: _get_class not permitted to load sdk Nov 27 02:47:56 localhost ceph-osd[32722]: _get_class not permitted to load test_remote_reads Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 crush map has features 288232575208783872, adjusting msgr requires for clients Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 crush map has features 288232575208783872 was 8705, adjusting msgr requires for mons Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 crush map has features 288232575208783872, adjusting msgr requires for osds Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 check_osdmap_features enabling on-disk ERASURE CODES compat feature Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 load_pgs Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 load_pgs opened 0 pgs Nov 27 02:47:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3[32718]: 2025-11-27T07:47:56.136+0000 7f01c7a80a80 -1 osd.3 0 log_to_monitors true Nov 27 02:47:56 localhost ceph-osd[32722]: osd.3 0 log_to_monitors true Nov 27 02:47:56 localhost systemd[1]: var-lib-containers-storage-overlay-9bb571fd4c853ae90a23ab5ce6aeafac05943d4693c4d575d50f93ab832f7e9e-merged.mount: Deactivated successfully. Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 0 maybe_override_max_osd_capacity_for_qos osd bench result - bandwidth (MiB/sec): 24.817 iops: 6353.225 elapsed_sec: 0.472 Nov 27 02:47:56 localhost ceph-osd[31760]: log_channel(cluster) log [WRN] : OSD bench result of 6353.225070 IOPS is not within the threshold limit range of 50.000000 IOPS and 500.000000 IOPS for osd.0. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's IOPS capacity using other benchmark tools (e.g. Fio) and then override osd_mclock_max_capacity_iops_[hdd|ssd]. Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 0 waiting for initial osdmap Nov 27 02:47:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0[31756]: 2025-11-27T07:47:56.538+0000 7f8f1c7f4640 -1 osd.0 0 waiting for initial osdmap Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 crush map has features 288514050185494528, adjusting msgr requires for clients Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 crush map has features 288514050185494528 was 288232575208792577, adjusting msgr requires for mons Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 crush map has features 3314932999778484224, adjusting msgr requires for osds Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 check_osdmap_features require_osd_release unknown -> reef Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 set_numa_affinity not setting numa affinity Nov 27 02:47:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-0[31756]: 2025-11-27T07:47:56.553+0000 7f8f17609640 -1 osd.0 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 10 _collect_metadata loop3: no unique device id for loop3: fallback method has no model nor serial Nov 27 02:47:56 localhost elated_villani[33245]: { Nov 27 02:47:56 localhost elated_villani[33245]: "042da33d-4846-4886-b930-3adda459b2df": { Nov 27 02:47:56 localhost elated_villani[33245]: "ceph_fsid": "e83f3b0c-4090-52df-95d4-ad9be8516692", Nov 27 02:47:56 localhost elated_villani[33245]: "device": "/dev/mapper/ceph_vg0-ceph_lv0", Nov 27 02:47:56 localhost elated_villani[33245]: "osd_id": 0, Nov 27 02:47:56 localhost elated_villani[33245]: "osd_uuid": "042da33d-4846-4886-b930-3adda459b2df", Nov 27 02:47:56 localhost elated_villani[33245]: "type": "bluestore" Nov 27 02:47:56 localhost elated_villani[33245]: }, Nov 27 02:47:56 localhost elated_villani[33245]: "c197799b-1f02-4d5b-a065-671596d4ef72": { Nov 27 02:47:56 localhost elated_villani[33245]: "ceph_fsid": "e83f3b0c-4090-52df-95d4-ad9be8516692", Nov 27 02:47:56 localhost elated_villani[33245]: "device": "/dev/mapper/ceph_vg1-ceph_lv1", Nov 27 02:47:56 localhost elated_villani[33245]: "osd_id": 3, Nov 27 02:47:56 localhost elated_villani[33245]: "osd_uuid": "c197799b-1f02-4d5b-a065-671596d4ef72", Nov 27 02:47:56 localhost elated_villani[33245]: "type": "bluestore" Nov 27 02:47:56 localhost elated_villani[33245]: } Nov 27 02:47:56 localhost elated_villani[33245]: } Nov 27 02:47:56 localhost systemd[1]: libpod-415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e.scope: Deactivated successfully. Nov 27 02:47:56 localhost podman[33048]: 2025-11-27 07:47:56.708308451 +0000 UTC m=+0.857549524 container died 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, vcs-type=git, release=553, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, name=rhceph, RELEASE=main, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, version=7, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, distribution-scope=public, description=Red Hat Ceph Storage 7) Nov 27 02:47:56 localhost systemd[1]: var-lib-containers-storage-overlay-4c03b0b894c24eeb55f59c6c056222fa2f980ce4307e1b3cf02ea1a26ea8643e-merged.mount: Deactivated successfully. Nov 27 02:47:56 localhost podman[33315]: 2025-11-27 07:47:56.794642326 +0000 UTC m=+0.074917866 container remove 415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elated_villani, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, ceph=True, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, description=Red Hat Ceph Storage 7, RELEASE=main, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, name=rhceph, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, version=7, vendor=Red Hat, Inc.) Nov 27 02:47:56 localhost systemd[1]: libpod-conmon-415eceef42ec78f467aa483bbf16e7d25eff0c9bbcc2b91eadc6a83d2c664d3e.scope: Deactivated successfully. Nov 27 02:47:56 localhost ceph-osd[31760]: osd.0 11 state: booting -> active Nov 27 02:47:57 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : purged_snaps scrub starts Nov 27 02:47:57 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : purged_snaps scrub ok Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 done with init, starting boot process Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 start_boot Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 maybe_override_options_for_qos osd_max_backfills set to 1 Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 maybe_override_options_for_qos osd_recovery_max_active set to 0 Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 maybe_override_options_for_qos osd_recovery_max_active_hdd set to 3 Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 maybe_override_options_for_qos osd_recovery_max_active_ssd set to 10 Nov 27 02:47:58 localhost ceph-osd[32722]: osd.3 0 bench count 12288000 bsize 4 KiB Nov 27 02:47:58 localhost podman[33439]: 2025-11-27 07:47:58.775934149 +0000 UTC m=+0.089135064 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, architecture=x86_64, CEPH_POINT_RELEASE=, RELEASE=main, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, maintainer=Guillaume Abrioux , release=553, GIT_BRANCH=main, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, name=rhceph) Nov 27 02:47:58 localhost podman[33439]: 2025-11-27 07:47:58.888814494 +0000 UTC m=+0.202015359 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, maintainer=Guillaume Abrioux , name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, release=553, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, distribution-scope=public, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, ceph=True, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, io.openshift.tags=rhceph ceph, vcs-type=git, io.openshift.expose-services=, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 02:48:00 localhost podman[33632]: Nov 27 02:48:00 localhost podman[33632]: 2025-11-27 07:48:00.852868667 +0000 UTC m=+0.079318961 container create af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, io.openshift.tags=rhceph ceph, RELEASE=main, GIT_BRANCH=main, CEPH_POINT_RELEASE=, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, ceph=True, GIT_CLEAN=True, com.redhat.component=rhceph-container, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, architecture=x86_64, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12) Nov 27 02:48:00 localhost podman[33632]: 2025-11-27 07:48:00.819302481 +0000 UTC m=+0.045752755 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:48:00 localhost systemd[1]: Started libpod-conmon-af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667.scope. Nov 27 02:48:00 localhost systemd[1]: Started libcrun container. Nov 27 02:48:00 localhost podman[33632]: 2025-11-27 07:48:00.976248761 +0000 UTC m=+0.202699025 container init af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, ceph=True, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main) Nov 27 02:48:00 localhost unruffled_banach[33647]: 167 167 Nov 27 02:48:00 localhost systemd[1]: libpod-af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667.scope: Deactivated successfully. Nov 27 02:48:01 localhost podman[33632]: 2025-11-27 07:48:01.017710803 +0000 UTC m=+0.244161107 container start af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, GIT_CLEAN=True, maintainer=Guillaume Abrioux , io.openshift.expose-services=, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, distribution-scope=public, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, architecture=x86_64) Nov 27 02:48:01 localhost podman[33632]: 2025-11-27 07:48:01.018343718 +0000 UTC m=+0.244794032 container attach af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , vcs-type=git, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, io.openshift.expose-services=, io.buildah.version=1.33.12, name=rhceph, GIT_CLEAN=True, version=7, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64) Nov 27 02:48:01 localhost podman[33632]: 2025-11-27 07:48:01.02220033 +0000 UTC m=+0.248650654 container died af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, version=7, ceph=True, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, distribution-scope=public, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , GIT_CLEAN=True, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc.) Nov 27 02:48:01 localhost systemd[1]: tmp-crun.1iuhjZ.mount: Deactivated successfully. Nov 27 02:48:01 localhost podman[33653]: 2025-11-27 07:48:01.125099268 +0000 UTC m=+0.117276370 container remove af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=unruffled_banach, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, distribution-scope=public, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , name=rhceph, RELEASE=main, GIT_BRANCH=main, io.buildah.version=1.33.12, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 02:48:01 localhost systemd[1]: libpod-conmon-af2e5e6837839b453342ddbf03dbc8ea9ccc77ee5ee26fa9936b93cc9c141667.scope: Deactivated successfully. Nov 27 02:48:01 localhost podman[33675]: Nov 27 02:48:01 localhost podman[33675]: 2025-11-27 07:48:01.333309573 +0000 UTC m=+0.080407877 container create 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, GIT_CLEAN=True, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, distribution-scope=public, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., ceph=True, version=7, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git) Nov 27 02:48:01 localhost systemd[1]: Started libpod-conmon-555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75.scope. Nov 27 02:48:01 localhost podman[33675]: 2025-11-27 07:48:01.292757741 +0000 UTC m=+0.039856145 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 02:48:01 localhost systemd[1]: Started libcrun container. Nov 27 02:48:01 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ac257309eee1455b4cafed54c8004cb2b846ff5bb1fdf36c4a5b77cbe982c668/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 02:48:01 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ac257309eee1455b4cafed54c8004cb2b846ff5bb1fdf36c4a5b77cbe982c668/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 02:48:01 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ac257309eee1455b4cafed54c8004cb2b846ff5bb1fdf36c4a5b77cbe982c668/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 02:48:01 localhost podman[33675]: 2025-11-27 07:48:01.440752678 +0000 UTC m=+0.187850972 container init 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, version=7, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., distribution-scope=public, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, CEPH_POINT_RELEASE=, release=553, architecture=x86_64, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 02:48:01 localhost podman[33675]: 2025-11-27 07:48:01.4497014 +0000 UTC m=+0.196799694 container start 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, architecture=x86_64, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, io.openshift.expose-services=, version=7, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_BRANCH=main, GIT_CLEAN=True) Nov 27 02:48:01 localhost podman[33675]: 2025-11-27 07:48:01.449896965 +0000 UTC m=+0.196995289 container attach 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, io.openshift.expose-services=, vendor=Red Hat, Inc., release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, RELEASE=main, distribution-scope=public, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, architecture=x86_64, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , name=rhceph, vcs-type=git, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 0 maybe_override_max_osd_capacity_for_qos osd bench result - bandwidth (MiB/sec): 21.611 iops: 5532.411 elapsed_sec: 0.542 Nov 27 02:48:01 localhost ceph-osd[32722]: log_channel(cluster) log [WRN] : OSD bench result of 5532.410501 IOPS is not within the threshold limit range of 50.000000 IOPS and 500.000000 IOPS for osd.3. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's IOPS capacity using other benchmark tools (e.g. Fio) and then override osd_mclock_max_capacity_iops_[hdd|ssd]. Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 0 waiting for initial osdmap Nov 27 02:48:01 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3[32718]: 2025-11-27T07:48:01.686+0000 7f01c39ff640 -1 osd.3 0 waiting for initial osdmap Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 crush map has features 288514051259236352, adjusting msgr requires for clients Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 crush map has features 288514051259236352 was 288232575208792577, adjusting msgr requires for mons Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 crush map has features 3314933000852226048, adjusting msgr requires for osds Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 check_osdmap_features require_osd_release unknown -> reef Nov 27 02:48:01 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-osd-3[32718]: 2025-11-27T07:48:01.708+0000 7f01bf029640 -1 osd.3 14 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 set_numa_affinity not setting numa affinity Nov 27 02:48:01 localhost ceph-osd[32722]: osd.3 14 _collect_metadata loop4: no unique device id for loop4: fallback method has no model nor serial Nov 27 02:48:01 localhost systemd[1]: var-lib-containers-storage-overlay-313c3c5a76e82a7703f5ea5474ba46ad7d517d9524f817595c91720fdd2f69fc-merged.mount: Deactivated successfully. Nov 27 02:48:02 localhost infallible_swartz[33691]: [ Nov 27 02:48:02 localhost infallible_swartz[33691]: { Nov 27 02:48:02 localhost infallible_swartz[33691]: "available": false, Nov 27 02:48:02 localhost infallible_swartz[33691]: "ceph_device": false, Nov 27 02:48:02 localhost infallible_swartz[33691]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 02:48:02 localhost infallible_swartz[33691]: "lsm_data": {}, Nov 27 02:48:02 localhost infallible_swartz[33691]: "lvs": [], Nov 27 02:48:02 localhost infallible_swartz[33691]: "path": "/dev/sr0", Nov 27 02:48:02 localhost infallible_swartz[33691]: "rejected_reasons": [ Nov 27 02:48:02 localhost infallible_swartz[33691]: "Has a FileSystem", Nov 27 02:48:02 localhost infallible_swartz[33691]: "Insufficient space (<5GB)" Nov 27 02:48:02 localhost infallible_swartz[33691]: ], Nov 27 02:48:02 localhost infallible_swartz[33691]: "sys_api": { Nov 27 02:48:02 localhost infallible_swartz[33691]: "actuators": null, Nov 27 02:48:02 localhost infallible_swartz[33691]: "device_nodes": "sr0", Nov 27 02:48:02 localhost infallible_swartz[33691]: "human_readable_size": "482.00 KB", Nov 27 02:48:02 localhost infallible_swartz[33691]: "id_bus": "ata", Nov 27 02:48:02 localhost infallible_swartz[33691]: "model": "QEMU DVD-ROM", Nov 27 02:48:02 localhost infallible_swartz[33691]: "nr_requests": "2", Nov 27 02:48:02 localhost infallible_swartz[33691]: "partitions": {}, Nov 27 02:48:02 localhost infallible_swartz[33691]: "path": "/dev/sr0", Nov 27 02:48:02 localhost infallible_swartz[33691]: "removable": "1", Nov 27 02:48:02 localhost infallible_swartz[33691]: "rev": "2.5+", Nov 27 02:48:02 localhost infallible_swartz[33691]: "ro": "0", Nov 27 02:48:02 localhost infallible_swartz[33691]: "rotational": "1", Nov 27 02:48:02 localhost infallible_swartz[33691]: "sas_address": "", Nov 27 02:48:02 localhost infallible_swartz[33691]: "sas_device_handle": "", Nov 27 02:48:02 localhost infallible_swartz[33691]: "scheduler_mode": "mq-deadline", Nov 27 02:48:02 localhost infallible_swartz[33691]: "sectors": 0, Nov 27 02:48:02 localhost infallible_swartz[33691]: "sectorsize": "2048", Nov 27 02:48:02 localhost infallible_swartz[33691]: "size": 493568.0, Nov 27 02:48:02 localhost infallible_swartz[33691]: "support_discard": "0", Nov 27 02:48:02 localhost infallible_swartz[33691]: "type": "disk", Nov 27 02:48:02 localhost infallible_swartz[33691]: "vendor": "QEMU" Nov 27 02:48:02 localhost infallible_swartz[33691]: } Nov 27 02:48:02 localhost infallible_swartz[33691]: } Nov 27 02:48:02 localhost infallible_swartz[33691]: ] Nov 27 02:48:02 localhost systemd[1]: libpod-555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75.scope: Deactivated successfully. Nov 27 02:48:02 localhost ceph-osd[32722]: osd.3 15 state: booting -> active Nov 27 02:48:02 localhost ceph-osd[32722]: osd.3 pg_epoch: 15 pg[1.0( empty local-lis/les=0/0 n=0 ec=12/12 lis/c=0/0 les/c/f=0/0/0 sis=15) [3,4,2] r=0 lpr=15 pi=[12,15)/0 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 02:48:02 localhost podman[35168]: 2025-11-27 07:48:02.374329262 +0000 UTC m=+0.058085688 container died 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, architecture=x86_64, version=7, vcs-type=git, release=553, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, name=rhceph, distribution-scope=public, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, RELEASE=main) Nov 27 02:48:02 localhost systemd[1]: var-lib-containers-storage-overlay-ac257309eee1455b4cafed54c8004cb2b846ff5bb1fdf36c4a5b77cbe982c668-merged.mount: Deactivated successfully. Nov 27 02:48:02 localhost podman[35168]: 2025-11-27 07:48:02.408218985 +0000 UTC m=+0.091975371 container remove 555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=infallible_swartz, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, architecture=x86_64, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, name=rhceph) Nov 27 02:48:02 localhost systemd[1]: libpod-conmon-555089defe1bbbdcf45e6bc86fcc99b40201928b0729d20f0c21976595f9ac75.scope: Deactivated successfully. Nov 27 02:48:03 localhost ceph-osd[31760]: osd.0 15 crush map has features 288514051259236352, adjusting msgr requires for clients Nov 27 02:48:03 localhost ceph-osd[31760]: osd.0 15 crush map has features 288514051259236352 was 288514050185503233, adjusting msgr requires for mons Nov 27 02:48:03 localhost ceph-osd[31760]: osd.0 15 crush map has features 3314933000852226048, adjusting msgr requires for osds Nov 27 02:48:03 localhost ceph-osd[32722]: osd.3 pg_epoch: 16 pg[1.0( empty local-lis/les=15/16 n=0 ec=12/12 lis/c=0/0 les/c/f=0/0/0 sis=15) [3,4,2] r=0 lpr=15 pi=[12,15)/0 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 02:48:11 localhost systemd[1]: tmp-crun.R2uNeJ.mount: Deactivated successfully. Nov 27 02:48:11 localhost podman[35295]: 2025-11-27 07:48:11.312569006 +0000 UTC m=+0.082038725 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, version=7, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, com.redhat.component=rhceph-container, distribution-scope=public, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, CEPH_POINT_RELEASE=, GIT_BRANCH=main, ceph=True) Nov 27 02:48:11 localhost podman[35295]: 2025-11-27 07:48:11.438326966 +0000 UTC m=+0.207796695 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-type=git, com.redhat.component=rhceph-container, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, GIT_CLEAN=True, CEPH_POINT_RELEASE=, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, version=7, RELEASE=main, distribution-scope=public) Nov 27 02:48:34 localhost systemd[26159]: Starting Mark boot as successful... Nov 27 02:48:34 localhost systemd[26159]: Finished Mark boot as successful. Nov 27 02:49:13 localhost podman[35476]: 2025-11-27 07:49:13.235772965 +0000 UTC m=+0.084575489 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, version=7, distribution-scope=public, CEPH_POINT_RELEASE=, RELEASE=main, GIT_CLEAN=True, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, name=rhceph, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc.) Nov 27 02:49:13 localhost podman[35476]: 2025-11-27 07:49:13.36793303 +0000 UTC m=+0.216735514 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, GIT_BRANCH=main, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, io.openshift.expose-services=, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., version=7, GIT_CLEAN=True, maintainer=Guillaume Abrioux , architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, release=553, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 02:49:23 localhost systemd[1]: session-13.scope: Deactivated successfully. Nov 27 02:49:23 localhost systemd[1]: session-13.scope: Consumed 20.743s CPU time. Nov 27 02:49:23 localhost systemd-logind[761]: Session 13 logged out. Waiting for processes to exit. Nov 27 02:49:23 localhost systemd-logind[761]: Removed session 13. Nov 27 02:51:34 localhost systemd[26159]: Created slice User Background Tasks Slice. Nov 27 02:51:34 localhost systemd[26159]: Starting Cleanup of User's Temporary Files and Directories... Nov 27 02:51:34 localhost systemd[26159]: Finished Cleanup of User's Temporary Files and Directories. Nov 27 02:51:35 localhost sshd[35771]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:52:47 localhost sshd[35849]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:52:47 localhost systemd-logind[761]: New session 27 of user zuul. Nov 27 02:52:47 localhost systemd[1]: Started Session 27 of User zuul. Nov 27 02:52:47 localhost python3[35897]: ansible-ansible.legacy.ping Invoked with data=pong Nov 27 02:52:49 localhost python3[35942]: ansible-setup Invoked with gather_subset=['!facter', '!ohai'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 02:52:50 localhost python3[35962]: ansible-user Invoked with name=tripleo-admin generate_ssh_key=False state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.localdomain update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None hidden=None seuser=None skeleton=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Nov 27 02:52:50 localhost python3[36018]: ansible-ansible.legacy.stat Invoked with path=/etc/sudoers.d/tripleo-admin follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:52:51 localhost python3[36061]: ansible-ansible.legacy.copy Invoked with dest=/etc/sudoers.d/tripleo-admin mode=288 owner=root group=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764229970.4825222-65903-20806206174076/source _original_basename=tmpigezwte9 follow=False checksum=b3e7ecdcc699d217c6b083a91b07208207813d93 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:52:51 localhost python3[36091]: ansible-file Invoked with path=/home/tripleo-admin state=directory owner=tripleo-admin group=tripleo-admin mode=448 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:52:52 localhost python3[36107]: ansible-file Invoked with path=/home/tripleo-admin/.ssh state=directory owner=tripleo-admin group=tripleo-admin mode=448 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:52:52 localhost python3[36123]: ansible-file Invoked with path=/home/tripleo-admin/.ssh/authorized_keys state=touch owner=tripleo-admin group=tripleo-admin mode=384 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:52:53 localhost python3[36139]: ansible-lineinfile Invoked with path=/home/tripleo-admin/.ssh/authorized_keys line=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC0lxMFmGTQVXeQKM/BF8qjBVvPBboTbQnDYFYfgRMDaSOn0SYW3wzC3mQeCkt7LixYs47VrmNV4ZkLMh1Y35QpuFU4tyterQasnSPlOWYxRzYQ+NYljwRxg6jZTX6HSZJmegB/aeobjHALfoMOfLN8hlwrIx0PTgae/7TN4GY+XKhWQCVvH5ImAjTUhoIR2hxkDsxy6O5gUuzAbhAUgO8RzMkKcCYGS8xFy/w4SzLBd1zPxD8rEJvptE7NGvaR4eCcPsgWlfnxStRJtYHjwH35dXHMP5l1P/TTcHAmpnc+tzdi0zYYXEDt1BTET0KkU96JU9XXjSBOmnhAUCws9joGXCRgF2M220OklI2HSp+CFfBz0yJ/L5QaN9gs9vKkgQbv5vB5hvR+gcHhsgFOT5PwFs/LjLijYpzhQKAooeRem88DQcEflZVJa43vw3KJdbq07YepCPFwn+jrV5HDWg37IxVPx6+PpC2PPJ9yEoFCuVbj6kl0m+P/yUm/ZNOvhvU= zuul-build-sshkey#012 regexp=Generated by TripleO state=present backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:52:53 localhost python3[36153]: ansible-ping Invoked with data=pong Nov 27 02:53:04 localhost sshd[36154]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:53:04 localhost systemd-logind[761]: New session 28 of user tripleo-admin. Nov 27 02:53:04 localhost systemd[1]: Created slice User Slice of UID 1003. Nov 27 02:53:04 localhost systemd[1]: Starting User Runtime Directory /run/user/1003... Nov 27 02:53:04 localhost systemd[1]: Finished User Runtime Directory /run/user/1003. Nov 27 02:53:04 localhost systemd[1]: Starting User Manager for UID 1003... Nov 27 02:53:05 localhost systemd[36158]: Queued start job for default target Main User Target. Nov 27 02:53:05 localhost systemd[36158]: Created slice User Application Slice. Nov 27 02:53:05 localhost systemd[36158]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 02:53:05 localhost systemd[36158]: Started Daily Cleanup of User's Temporary Directories. Nov 27 02:53:05 localhost systemd[36158]: Reached target Paths. Nov 27 02:53:05 localhost systemd[36158]: Reached target Timers. Nov 27 02:53:05 localhost systemd[36158]: Starting D-Bus User Message Bus Socket... Nov 27 02:53:05 localhost systemd[36158]: Starting Create User's Volatile Files and Directories... Nov 27 02:53:05 localhost systemd[36158]: Listening on D-Bus User Message Bus Socket. Nov 27 02:53:05 localhost systemd[36158]: Finished Create User's Volatile Files and Directories. Nov 27 02:53:05 localhost systemd[36158]: Reached target Sockets. Nov 27 02:53:05 localhost systemd[36158]: Reached target Basic System. Nov 27 02:53:05 localhost systemd[36158]: Reached target Main User Target. Nov 27 02:53:05 localhost systemd[36158]: Startup finished in 122ms. Nov 27 02:53:05 localhost systemd[1]: Started User Manager for UID 1003. Nov 27 02:53:05 localhost systemd[1]: Started Session 28 of User tripleo-admin. Nov 27 02:53:05 localhost python3[36220]: ansible-ansible.legacy.setup Invoked with gather_subset=['!all', 'min'] gather_timeout=45 filter=[] fact_path=/etc/ansible/facts.d Nov 27 02:53:10 localhost python3[36240]: ansible-selinux Invoked with policy=targeted state=enforcing configfile=/etc/selinux/config Nov 27 02:53:11 localhost python3[36256]: ansible-tempfile Invoked with state=file suffix=tmphosts prefix=ansible. path=None Nov 27 02:53:11 localhost python3[36304]: ansible-ansible.legacy.copy Invoked with remote_src=True src=/etc/hosts dest=/tmp/ansible.o61p1s15tmphosts mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:53:12 localhost python3[36334]: ansible-blockinfile Invoked with state=absent path=/tmp/ansible.o61p1s15tmphosts block= marker=# {mark} marker_begin=HEAT_HOSTS_START - Do not edit manually within this section! marker_end=HEAT_HOSTS_END create=False backup=False unsafe_writes=False insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:53:13 localhost python3[36350]: ansible-blockinfile Invoked with create=True path=/tmp/ansible.o61p1s15tmphosts insertbefore=BOF block=172.17.0.106 np0005537444.localdomain np0005537444#012172.18.0.106 np0005537444.storage.localdomain np0005537444.storage#012172.20.0.106 np0005537444.storagemgmt.localdomain np0005537444.storagemgmt#012172.17.0.106 np0005537444.internalapi.localdomain np0005537444.internalapi#012172.19.0.106 np0005537444.tenant.localdomain np0005537444.tenant#012192.168.122.106 np0005537444.ctlplane.localdomain np0005537444.ctlplane#012172.17.0.107 np0005537445.localdomain np0005537445#012172.18.0.107 np0005537445.storage.localdomain np0005537445.storage#012172.20.0.107 np0005537445.storagemgmt.localdomain np0005537445.storagemgmt#012172.17.0.107 np0005537445.internalapi.localdomain np0005537445.internalapi#012172.19.0.107 np0005537445.tenant.localdomain np0005537445.tenant#012192.168.122.107 np0005537445.ctlplane.localdomain np0005537445.ctlplane#012172.17.0.108 np0005537446.localdomain np0005537446#012172.18.0.108 np0005537446.storage.localdomain np0005537446.storage#012172.20.0.108 np0005537446.storagemgmt.localdomain np0005537446.storagemgmt#012172.17.0.108 np0005537446.internalapi.localdomain np0005537446.internalapi#012172.19.0.108 np0005537446.tenant.localdomain np0005537446.tenant#012192.168.122.108 np0005537446.ctlplane.localdomain np0005537446.ctlplane#012172.17.0.103 np0005537441.localdomain np0005537441#012172.18.0.103 np0005537441.storage.localdomain np0005537441.storage#012172.20.0.103 np0005537441.storagemgmt.localdomain np0005537441.storagemgmt#012172.17.0.103 np0005537441.internalapi.localdomain np0005537441.internalapi#012172.19.0.103 np0005537441.tenant.localdomain np0005537441.tenant#012192.168.122.103 np0005537441.ctlplane.localdomain np0005537441.ctlplane#012172.17.0.104 np0005537442.localdomain np0005537442#012172.18.0.104 np0005537442.storage.localdomain np0005537442.storage#012172.20.0.104 np0005537442.storagemgmt.localdomain np0005537442.storagemgmt#012172.17.0.104 np0005537442.internalapi.localdomain np0005537442.internalapi#012172.19.0.104 np0005537442.tenant.localdomain np0005537442.tenant#012192.168.122.104 np0005537442.ctlplane.localdomain np0005537442.ctlplane#012172.17.0.105 np0005537443.localdomain np0005537443#012172.18.0.105 np0005537443.storage.localdomain np0005537443.storage#012172.20.0.105 np0005537443.storagemgmt.localdomain np0005537443.storagemgmt#012172.17.0.105 np0005537443.internalapi.localdomain np0005537443.internalapi#012172.19.0.105 np0005537443.tenant.localdomain np0005537443.tenant#012192.168.122.105 np0005537443.ctlplane.localdomain np0005537443.ctlplane#012#012192.168.122.100 undercloud.ctlplane.localdomain undercloud.ctlplane#012192.168.122.99 overcloud.ctlplane.localdomain#012172.18.0.158 overcloud.storage.localdomain#012172.20.0.184 overcloud.storagemgmt.localdomain#012172.17.0.185 overcloud.internalapi.localdomain#012172.21.0.212 overcloud.localdomain#012 marker=# {mark} marker_begin=START_HOST_ENTRIES_FOR_STACK: overcloud marker_end=END_HOST_ENTRIES_FOR_STACK: overcloud state=present backup=False unsafe_writes=False insertafter=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:53:14 localhost python3[36366]: ansible-ansible.legacy.command Invoked with _raw_params=cp "/tmp/ansible.o61p1s15tmphosts" "/etc/hosts" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:53:14 localhost python3[36383]: ansible-file Invoked with path=/tmp/ansible.o61p1s15tmphosts state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:53:15 localhost python3[36399]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q --whatprovides rhosp-release _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:53:16 localhost python3[36416]: ansible-ansible.legacy.dnf Invoked with name=['rhosp-release'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:53:20 localhost python3[36497]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q --whatprovides driverctl lvm2 jq nftables openvswitch openstack-heat-agents openstack-selinux os-net-config python3-libselinux python3-pyyaml puppet-tripleo rsync tmpwatch sysstat iproute-tc _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:53:21 localhost python3[36514]: ansible-ansible.legacy.dnf Invoked with name=['driverctl', 'lvm2', 'jq', 'nftables', 'openvswitch', 'openstack-heat-agents', 'openstack-selinux', 'os-net-config', 'python3-libselinux', 'python3-pyyaml', 'puppet-tripleo', 'rsync', 'tmpwatch', 'sysstat', 'iproute-tc'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:54:31 localhost kernel: SELinux: Converting 2699 SID table entries... Nov 27 02:54:31 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:54:31 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:54:31 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:54:31 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:54:31 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:54:31 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:54:31 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:54:31 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=6 res=1 Nov 27 02:54:32 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:54:32 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:54:32 localhost systemd[1]: Reloading. Nov 27 02:54:32 localhost systemd-sysv-generator[38646]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:54:32 localhost systemd-rc-local-generator[38642]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:54:32 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:54:32 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:54:32 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:54:32 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:54:32 localhost systemd[1]: run-rdff5f1a820e340eb860ad4ece7a4d48f.service: Deactivated successfully. Nov 27 02:54:33 localhost python3[39089]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -V driverctl lvm2 jq nftables openvswitch openstack-heat-agents openstack-selinux os-net-config python3-libselinux python3-pyyaml puppet-tripleo rsync tmpwatch sysstat iproute-tc _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:35 localhost python3[39228]: ansible-ansible.legacy.systemd Invoked with name=openvswitch enabled=True state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:54:35 localhost systemd[1]: Reloading. Nov 27 02:54:35 localhost systemd-sysv-generator[39260]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:54:35 localhost systemd-rc-local-generator[39255]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:54:35 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:54:37 localhost python3[39282]: ansible-file Invoked with path=/var/lib/heat-config/tripleo-config-download state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:37 localhost python3[39298]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q --whatprovides openstack-network-scripts _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:38 localhost python3[39315]: ansible-systemd Invoked with name=NetworkManager enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 02:54:38 localhost python3[39333]: ansible-ini_file Invoked with path=/etc/NetworkManager/NetworkManager.conf state=present no_extra_spaces=True section=main option=dns value=none backup=True exclusive=True allow_no_value=False create=True unsafe_writes=False values=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:39 localhost python3[39351]: ansible-ini_file Invoked with path=/etc/NetworkManager/NetworkManager.conf state=present no_extra_spaces=True section=main option=rc-manager value=unmanaged backup=True exclusive=True allow_no_value=False create=True unsafe_writes=False values=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:39 localhost python3[39369]: ansible-ansible.legacy.systemd Invoked with name=NetworkManager state=reloaded daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:54:39 localhost systemd[1]: Reloading Network Manager... Nov 27 02:54:39 localhost NetworkManager[5971]: [1764230079.8034] audit: op="reload" arg="0" pid=39372 uid=0 result="success" Nov 27 02:54:39 localhost NetworkManager[5971]: [1764230079.8042] config: signal: SIGHUP,config-files,values,values-user,no-auto-default,dns-mode,rc-manager (/etc/NetworkManager/NetworkManager.conf (lib: 00-server.conf) (run: 15-carrier-timeout.conf)) Nov 27 02:54:39 localhost NetworkManager[5971]: [1764230079.8043] dns-mgr: init: dns=none,systemd-resolved rc-manager=unmanaged Nov 27 02:54:39 localhost systemd[1]: Reloaded Network Manager. Nov 27 02:54:40 localhost python3[39388]: ansible-ansible.legacy.command Invoked with _raw_params=ln -f -s /usr/share/openstack-puppet/modules/* /etc/puppet/modules/ _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:40 localhost python3[39405]: ansible-stat Invoked with path=/usr/bin/ansible-playbook follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:54:41 localhost python3[39423]: ansible-stat Invoked with path=/usr/bin/ansible-playbook-3 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:54:41 localhost python3[39439]: ansible-file Invoked with state=link src=/usr/bin/ansible-playbook path=/usr/bin/ansible-playbook-3 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:42 localhost python3[39455]: ansible-tempfile Invoked with state=file prefix=ansible. suffix= path=None Nov 27 02:54:42 localhost python3[39471]: ansible-stat Invoked with path=/etc/ssh/ssh_known_hosts follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:54:43 localhost python3[39487]: ansible-blockinfile Invoked with path=/tmp/ansible.z5szjd6e block=[192.168.122.106]*,[np0005537444.ctlplane.localdomain]*,[172.17.0.106]*,[np0005537444.internalapi.localdomain]*,[172.18.0.106]*,[np0005537444.storage.localdomain]*,[172.20.0.106]*,[np0005537444.storagemgmt.localdomain]*,[172.19.0.106]*,[np0005537444.tenant.localdomain]*,[np0005537444.localdomain]*,[np0005537444]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVQnNmX1+68w+ZbvmHMSJP3G6Gp5eoU/UG1TPEqifZH8HzinCXMNkeyCf3BVotH4CvjVs1UyzcJAGbOj+4N9NMuosvJcf7WTfQSxAkJB9Dz3gV9yg/LAScQsvVi+t9BxPkiPLrjs/JSyy4K6D0LQ8TLlXrhcrwVTHQvGKPvJqlD2RO+kLhWz/jR7en1FPLamRhvW3w6dAWI+nYsKUUGDdu6lsb2nHwAEyIHnLp+IeCbeM/apPU9DSh6ERNERW8QAqmBIDeEp3C1sNEtiBMVtj9fAH7SqTI/fRnJTD1LP6ziSGlVXgFJdkCs4lPmNuOEagV9DDO7l4eD0kzAJ1o/a8uvUDkRBHjrRExUxVeaOH2hNyNY5sZ0Cw9CZXScubJJkxlKbMJuzHw3ZpvoECi1yaJVCYiQ/1JznRtXbEHwBstIiCLmrUtEiMoZ7oH45v8B9yHNHdtuxZQHYybx+4JFrv2uqtUEhJ2QmWGqTIQGAA0ITLL6NznaorLZDpDkMuDBXs=#012[192.168.122.107]*,[np0005537445.ctlplane.localdomain]*,[172.17.0.107]*,[np0005537445.internalapi.localdomain]*,[172.18.0.107]*,[np0005537445.storage.localdomain]*,[172.20.0.107]*,[np0005537445.storagemgmt.localdomain]*,[172.19.0.107]*,[np0005537445.tenant.localdomain]*,[np0005537445.localdomain]*,[np0005537445]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1/piBT2PI0lEnlBA9RLkYttkVms78XEr2pK4jrGzXLJsjY5YhAhpZjABbX6IrEXb0jtLr9fu2OK3NtHKlrjNb/7TkxYw4YuiZWvdYFf4s6Q4FTcCO6oLrFLcEgu5Rzzmv0bCZSCXSH7ZrDLbW3KwkiwaTCUIu9j3GxfcmFYjiTzDPQTsskQc06ojYDvAc0mflNdibjhZHmD73Yu4NM5EtulvOe7cIAF5fk6E8y41QtFlDZVKun2C2copAdORw4q1KMoxBeHLUDgGQniBbhgct43//SHZpvNiSjsvvV1CyEotNjbAd1nyuAiHMDy9mON8NSR4hJfrBRXNDNmenoJNQErYdcIJQZwiXMZEGzVgK2LPcTnlyGKMwZdailMhaZ7vqxYN9TxRRUxsopDhhk4ekYF9FtxBbt2KfRzjqQOkmiiNaYAiBdUCd/ogfYfRl54F0j/FGmrljjHOw0oPcQVd/cZiwZrgjwnyoMahxDpiCRE21ptokmRC8f24fhQXEJ80=#012[192.168.122.108]*,[np0005537446.ctlplane.localdomain]*,[172.17.0.108]*,[np0005537446.internalapi.localdomain]*,[172.18.0.108]*,[np0005537446.storage.localdomain]*,[172.20.0.108]*,[np0005537446.storagemgmt.localdomain]*,[172.19.0.108]*,[np0005537446.tenant.localdomain]*,[np0005537446.localdomain]*,[np0005537446]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZeWh6gZHJpy+jtMRG/XGdhy9SYywGs3q0MIBAiWgx7/hzkaCQxjMMvFVu6G3hROUk6sKQGhF8WPWsE+0AjjsrUl2xCmaViqBT1R0jfZC2U19+mKEhJVUwet4N5BMXRm5YziIc6/BNWGrQ+7gPl1K94zxP8dR4FoSEhlGn2HV4p1iKBt5Z93ZjsRERdN4vUP78W0UfsxStluowzwz5Nu++6mg/D64q1n+arRV9mW3/rHAU1FHGhhK820Y7T3PQgxdBx9GjP7EwDoWITYATGCfWS1a/fMKs5faBezOWJ5dEo7IOUZILjEuOvEJQMCgEsPwUd8G3q0T0uq1BMvdjp8iRhP7hp4ggEB/ciIP3zgnw0g/fsbvQ62pqKs7yWr1jt17deQInxZK+U6oq3Np67FCX8kfJxaAOB98JarjtuJVhrAWhPCRvldwkuqJ7QliNlo4h9ceQAvZNoYMSPOVDh1UwstPXUxyp0l5AQBGqU8jzi3Mt9S0gFTkG1FsGcwWGL6U=#012[192.168.122.103]*,[np0005537441.ctlplane.localdomain]*,[172.17.0.103]*,[np0005537441.internalapi.localdomain]*,[172.18.0.103]*,[np0005537441.storage.localdomain]*,[172.20.0.103]*,[np0005537441.storagemgmt.localdomain]*,[172.19.0.103]*,[np0005537441.tenant.localdomain]*,[np0005537441.localdomain]*,[np0005537441]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDWRAei0LGbei1vF4gcoA5zoYcENUUXbT+RNxRrzeTBr84EZ6DQ6p9T5XNKUcEoe7tPEv7QsuZekWJ50wtUWSZRlQCeG+B/cTNlvRx3tfOciPFmCM7x149Oo3xTUEk+TxUgg+U31VxWmmVLod2YnhscV+n5AOSoCWF6F8PqLVJI1q7TQYzmKK5C4QkD7eNwDBkMDHaeWMnXOdkw025dKxRCzE7BttEEmXznEskx6l9MQR432QsxjGBhUNCvHYwG6/fm31ELhOX1BITpI5DzQu3bjnxHLTSgpmvP5Y6AgXn/Snh0394S4PeMDEzK2ivAMuC0mG/dPet7nDKlY8QZrsd8TH4a9REwvZ3DKXbYx5odMeB357AyW1CsKniltq/Icq4sOES9STUZEowOlFZfcqc9Oh7o89+vEvORNNKcLMq91qDTlBFBFDjsPYvkCzhCSlxUCZ1AICp6prURQDrfmwM881TUKqMZsgnOmXjz/TWlE5e/XW5fmXrRTxMFljrpYgk=#012[192.168.122.104]*,[np0005537442.ctlplane.localdomain]*,[172.17.0.104]*,[np0005537442.internalapi.localdomain]*,[172.18.0.104]*,[np0005537442.storage.localdomain]*,[172.20.0.104]*,[np0005537442.storagemgmt.localdomain]*,[172.19.0.104]*,[np0005537442.tenant.localdomain]*,[np0005537442.localdomain]*,[np0005537442]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDZ6+t2RQvYi/f60sV6qGAF60A1plHYLpQn7dD2IhG+Czr5OEjxITemq/7OJS7h+Ne9DU0X2SWSz0GMHkvJzD8MWBUTqEiZSsKGpTU5TzJgBbtQh4a/LJPLTMVeWh1bGNqmhfMab+H676ZaRem42hcIaqksBkctHcH+vYEugPUULNxUUYotCsiDPtQcGY4sB0PAH+TXLOLsdSLlWFAn8SaZLGixyiXxQzRJBBewu5KDE3jRFntZPNjAZUmgOKUNHTkNCwTCURrVkjOqTx2h4cbJFLTgH5cua9N1SClN7wCHHd9nOZKZJokrHltXDSxGfdSMM8LnQ0vl8/tkQ0mZxST1LVzUcUeTW+kNz44ygmPbBpR3F4sWj4zM+miT/2EarGnH87pJ3NeiNNp2ZEBx9uoowFcdECFaIBSf8+29k3AdB5tFYsGzkKQ81cZjNsC8bjDU3LNufSmtl922xHJuXclY1wsdLHJjerf5NNBh4V/5DzzBOC8tQuuJcIlUBd8mfX0=#012[192.168.122.105]*,[np0005537443.ctlplane.localdomain]*,[172.17.0.105]*,[np0005537443.internalapi.localdomain]*,[172.18.0.105]*,[np0005537443.storage.localdomain]*,[172.20.0.105]*,[np0005537443.storagemgmt.localdomain]*,[172.19.0.105]*,[np0005537443.tenant.localdomain]*,[np0005537443.localdomain]*,[np0005537443]* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+TzqlkhU2yG8BpCyg9Jf4bhGZbiRnL+Jc1FYzYFLOntd5BBvaw6hhYNzDc/mKZeP0tlyeFmf0aA+wYpdWrTwIgTPqGx4UhnipnEPZSEFfJn8o0fzHqgZ5h22y7XkklotRytWOPX87vHDdeYDciC1tnzCEdxTJMYwBzM41MaEXcO+Y517qGhLL8ftIsAjgvw9GK5z6h1iEA66zTONVmi/I6HZhjaRkhUmhi4oQ39ipF6wSAd79aQikmsuUq2NBHKGbvonjGPTI6ni5CmjVWa4dTpua5+pCHgJAjYmrD+nZvAPxHPdz2Z4H9rlYsmC2An0sOXROsGuHEeQzoSPmVbXTH9m865cib1g3r+C6usVsK4pVA6pS8TEXkAEJ61rQWYLdxykz52kixRP242TnVJU+znQgKsgJWcD7997B6BNLZwgB2zSXz+uJRVgQvUQ1ktXL/TtVO+OUrnrBEUzIxZju/ukUoFQyWR8IXwX6U9d0ozh7cXKp/bZ2ah9U0EWnLTc=#012 create=True state=present marker=# {mark} ANSIBLE MANAGED BLOCK backup=False marker_begin=BEGIN marker_end=END unsafe_writes=False insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:44 localhost python3[39503]: ansible-ansible.legacy.command Invoked with _raw_params=cat '/tmp/ansible.z5szjd6e' > /etc/ssh/ssh_known_hosts _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:44 localhost python3[39521]: ansible-file Invoked with path=/tmp/ansible.z5szjd6e state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:54:45 localhost python3[39537]: ansible-file Invoked with path=/var/log/journal state=directory mode=0750 owner=root group=root setype=var_log_t recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 02:54:45 localhost python3[39553]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-active cloud-init.service || systemctl is-enabled cloud-init.service _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:46 localhost python3[39571]: ansible-ansible.legacy.command Invoked with _raw_params=cat /proc/cmdline | grep -q cloud-init=disabled _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:46 localhost python3[39590]: ansible-community.general.cloud_init_data_facts Invoked with filter=status Nov 27 02:54:49 localhost python3[39727]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q --whatprovides tuned tuned-profiles-cpu-partitioning _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:49 localhost python3[39744]: ansible-ansible.legacy.dnf Invoked with name=['tuned', 'tuned-profiles-cpu-partitioning'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:54:52 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:54:53 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:54:53 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:54:53 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:54:53 localhost systemd[1]: Reloading. Nov 27 02:54:53 localhost systemd-rc-local-generator[39807]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:54:53 localhost systemd-sysv-generator[39811]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:54:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:54:53 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:54:53 localhost systemd[1]: Stopping Dynamic System Tuning Daemon... Nov 27 02:54:53 localhost systemd[1]: tuned.service: Deactivated successfully. Nov 27 02:54:53 localhost systemd[1]: Stopped Dynamic System Tuning Daemon. Nov 27 02:54:53 localhost systemd[1]: tuned.service: Consumed 1.723s CPU time. Nov 27 02:54:53 localhost systemd[1]: Starting Dynamic System Tuning Daemon... Nov 27 02:54:53 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:54:53 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:54:53 localhost systemd[1]: run-r27c3e0c5288d4f5f821635fbddcc8b30.service: Deactivated successfully. Nov 27 02:54:55 localhost systemd[1]: Started Dynamic System Tuning Daemon. Nov 27 02:54:55 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:54:55 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:54:55 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:54:55 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:54:55 localhost systemd[1]: run-raaa28c2779e14d51965f2c402a1d838a.service: Deactivated successfully. Nov 27 02:54:56 localhost python3[40182]: ansible-systemd Invoked with name=tuned state=restarted enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:54:56 localhost systemd[1]: Stopping Dynamic System Tuning Daemon... Nov 27 02:54:56 localhost systemd[1]: tuned.service: Deactivated successfully. Nov 27 02:54:56 localhost systemd[1]: Stopped Dynamic System Tuning Daemon. Nov 27 02:54:56 localhost systemd[1]: Starting Dynamic System Tuning Daemon... Nov 27 02:54:57 localhost systemd[1]: Started Dynamic System Tuning Daemon. Nov 27 02:54:58 localhost python3[40377]: ansible-ansible.legacy.command Invoked with _raw_params=which tuned-adm _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:54:58 localhost python3[40394]: ansible-slurp Invoked with src=/etc/tuned/active_profile Nov 27 02:54:59 localhost python3[40410]: ansible-stat Invoked with path=/etc/tuned/throughput-performance-variables.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:54:59 localhost python3[40426]: ansible-ansible.legacy.command Invoked with _raw_params=tuned-adm profile throughput-performance _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:01 localhost python3[40446]: ansible-ansible.legacy.command Invoked with _raw_params=cat /proc/cmdline _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:02 localhost python3[40463]: ansible-stat Invoked with path=/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova/nova.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:55:05 localhost python3[40479]: ansible-replace Invoked with regexp=TRIPLEO_HEAT_TEMPLATE_KERNEL_ARGS dest=/etc/default/grub replace= path=/etc/default/grub backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:10 localhost python3[40495]: ansible-file Invoked with path=/etc/puppet/hieradata state=directory mode=448 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:10 localhost python3[40543]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hiera.yaml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:10 localhost python3[40588]: ansible-ansible.legacy.copy Invoked with mode=384 dest=/etc/puppet/hiera.yaml src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230110.2652128-70442-198313569592538/source _original_basename=tmpoc8epww9 follow=False checksum=aaf3699defba931d532f4955ae152f505046749a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:11 localhost python3[40618]: ansible-file Invoked with src=/etc/puppet/hiera.yaml dest=/etc/hiera.yaml state=link force=True path=/etc/hiera.yaml recurse=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:12 localhost python3[40666]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/all_nodes.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:12 localhost python3[40709]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230111.8653965-70538-188374912732042/source dest=/etc/puppet/hieradata/all_nodes.json _original_basename=overcloud.json follow=False checksum=a879ceb0a38cf9fbaae711346fd7314213d8dfe3 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:13 localhost python3[40771]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/bootstrap_node.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:13 localhost python3[40814]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230112.7654195-70597-252602244598328/source dest=/etc/puppet/hieradata/bootstrap_node.json mode=None follow=False _original_basename=bootstrap_node.j2 checksum=28e1692230a0b02660ab24a7a8d42d6368d4d36d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:14 localhost python3[40876]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/vip_data.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:14 localhost python3[40919]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230113.6592944-70597-200086587785065/source dest=/etc/puppet/hieradata/vip_data.json mode=None follow=False _original_basename=vip_data.j2 checksum=cb1409ab025f79dfe913ff6c91797f477bd6ef67 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:14 localhost python3[40981]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/net_ip_map.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:15 localhost python3[41024]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230114.5257714-70597-23530328238986/source dest=/etc/puppet/hieradata/net_ip_map.json mode=None follow=False _original_basename=net_ip_map.j2 checksum=1bd75eeb71ad8a06f7ad5bd2e02e7279e09e867f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:15 localhost python3[41086]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/cloud_domain.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:16 localhost python3[41129]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230115.454761-70597-22844789252683/source dest=/etc/puppet/hieradata/cloud_domain.json mode=None follow=False _original_basename=cloud_domain.j2 checksum=5dd835a63e6a03d74797c2e2eadf4bea1cecd9d9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:16 localhost python3[41191]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/fqdn.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:16 localhost python3[41234]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230116.2569876-70597-85156071448166/source dest=/etc/puppet/hieradata/fqdn.json mode=None follow=False _original_basename=fqdn.j2 checksum=b9c07d44b18c05328321c1ee33c0f9afd31f5e19 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:17 localhost python3[41296]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/service_names.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:17 localhost python3[41339]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230117.1114166-70597-87907872601994/source dest=/etc/puppet/hieradata/service_names.json mode=None follow=False _original_basename=service_names.j2 checksum=ff586b96402d8ae133745cf06f17e772b2f22d52 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:18 localhost python3[41401]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/service_configs.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:18 localhost python3[41444]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230117.9904103-70597-32478513911796/source dest=/etc/puppet/hieradata/service_configs.json mode=None follow=False _original_basename=service_configs.j2 checksum=f4969c6b6333a51d4040d04d97254aeba09bf0db backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:19 localhost python3[41506]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/extraconfig.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:19 localhost python3[41549]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230118.8961086-70597-103495488533118/source dest=/etc/puppet/hieradata/extraconfig.json mode=None follow=False _original_basename=extraconfig.j2 checksum=5f36b2ea290645ee34d943220a14b54ee5ea5be5 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:20 localhost python3[41611]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/role_extraconfig.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:20 localhost python3[41654]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230119.714675-70597-113949372946052/source dest=/etc/puppet/hieradata/role_extraconfig.json mode=None follow=False _original_basename=role_extraconfig.j2 checksum=34875968bf996542162e620523f9dcfb3deac331 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:20 localhost python3[41716]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/ovn_chassis_mac_map.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:21 localhost python3[41759]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230120.5511782-70597-45484995075767/source dest=/etc/puppet/hieradata/ovn_chassis_mac_map.json mode=None follow=False _original_basename=ovn_chassis_mac_map.j2 checksum=cf0f9c716d9e566af094ed1faac23e8b18804ff5 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:21 localhost python3[41789]: ansible-stat Invoked with path={'src': '/etc/puppet/hieradata/ansible_managed.json'} follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:55:22 localhost python3[41837]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/ansible_managed.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:55:23 localhost python3[41910]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/ansible_managed.json owner=root group=root mode=0644 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230122.2961526-71394-178867597320423/source _original_basename=tmpzqn9q_k4 follow=False checksum=bf21a9e8fbc5a3846fb05b4fa0859e0917b2202f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:55:27 localhost python3[41986]: ansible-setup Invoked with gather_subset=['!all', '!min', 'network'] filter=['ansible_default_ipv4'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 02:55:28 localhost python3[42047]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 38.102.83.1 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:32 localhost python3[42064]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 192.168.122.10 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:34 localhost systemd[36158]: Starting Mark boot as successful... Nov 27 02:55:34 localhost systemd[36158]: Finished Mark boot as successful. Nov 27 02:55:38 localhost python3[42082]: ansible-ansible.legacy.command Invoked with _raw_params=INT=$(ip ro get 192.168.122.106 | head -1 | sed -nr "s/.* dev (\w+) .*/\1/p")#012MTU=$(cat /sys/class/net/${INT}/mtu 2>/dev/null || echo "0")#012echo "$INT $MTU"#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:38 localhost python3[42105]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 192.168.122.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:43 localhost python3[42122]: ansible-ansible.legacy.command Invoked with _raw_params=INT=$(ip ro get 172.18.0.106 | head -1 | sed -nr "s/.* dev (\w+) .*/\1/p")#012MTU=$(cat /sys/class/net/${INT}/mtu 2>/dev/null || echo "0")#012echo "$INT $MTU"#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:43 localhost python3[42145]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 172.18.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:48 localhost python3[42162]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -s 1472 -c 5 172.18.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:52 localhost python3[42179]: ansible-ansible.legacy.command Invoked with _raw_params=INT=$(ip ro get 172.20.0.106 | head -1 | sed -nr "s/.* dev (\w+) .*/\1/p")#012MTU=$(cat /sys/class/net/${INT}/mtu 2>/dev/null || echo "0")#012echo "$INT $MTU"#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:53 localhost python3[42202]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 172.20.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:55:57 localhost python3[42219]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -s 1472 -c 5 172.20.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:02 localhost python3[42236]: ansible-ansible.legacy.command Invoked with _raw_params=INT=$(ip ro get 172.17.0.106 | head -1 | sed -nr "s/.* dev (\w+) .*/\1/p")#012MTU=$(cat /sys/class/net/${INT}/mtu 2>/dev/null || echo "0")#012echo "$INT $MTU"#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:02 localhost python3[42259]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 172.17.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:07 localhost python3[42276]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -s 1472 -c 5 172.17.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:11 localhost python3[42293]: ansible-ansible.legacy.command Invoked with _raw_params=INT=$(ip ro get 172.19.0.106 | head -1 | sed -nr "s/.* dev (\w+) .*/\1/p")#012MTU=$(cat /sys/class/net/${INT}/mtu 2>/dev/null || echo "0")#012echo "$INT $MTU"#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:12 localhost python3[42316]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -c 5 172.19.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:16 localhost python3[42333]: ansible-ansible.legacy.command Invoked with _raw_params=ping -w 10 -s 1472 -c 5 172.19.0.106 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:21 localhost python3[42350]: ansible-file Invoked with path=/etc/puppet/hieradata state=directory mode=448 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:22 localhost python3[42398]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hiera.yaml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:22 localhost python3[42416]: ansible-ansible.legacy.file Invoked with mode=384 dest=/etc/puppet/hiera.yaml _original_basename=tmp3osv2ng0 recurse=False state=file path=/etc/puppet/hiera.yaml force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:23 localhost python3[42446]: ansible-file Invoked with src=/etc/puppet/hiera.yaml dest=/etc/hiera.yaml state=link force=True path=/etc/hiera.yaml recurse=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:23 localhost python3[42494]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/all_nodes.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:24 localhost python3[42512]: ansible-ansible.legacy.file Invoked with dest=/etc/puppet/hieradata/all_nodes.json _original_basename=overcloud.json recurse=False state=file path=/etc/puppet/hieradata/all_nodes.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:24 localhost python3[42619]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/bootstrap_node.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:24 localhost python3[42654]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/bootstrap_node.json _original_basename=bootstrap_node.j2 recurse=False state=file path=/etc/puppet/hieradata/bootstrap_node.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:25 localhost python3[42716]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/vip_data.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:25 localhost python3[42734]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/vip_data.json _original_basename=vip_data.j2 recurse=False state=file path=/etc/puppet/hieradata/vip_data.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:26 localhost python3[42796]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/net_ip_map.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:26 localhost python3[42814]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/net_ip_map.json _original_basename=net_ip_map.j2 recurse=False state=file path=/etc/puppet/hieradata/net_ip_map.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:26 localhost python3[42876]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/cloud_domain.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:27 localhost python3[42909]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/cloud_domain.json _original_basename=cloud_domain.j2 recurse=False state=file path=/etc/puppet/hieradata/cloud_domain.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:27 localhost python3[42971]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/fqdn.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:27 localhost python3[42989]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/fqdn.json _original_basename=fqdn.j2 recurse=False state=file path=/etc/puppet/hieradata/fqdn.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:28 localhost python3[43051]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/service_names.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:28 localhost python3[43069]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/service_names.json _original_basename=service_names.j2 recurse=False state=file path=/etc/puppet/hieradata/service_names.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:29 localhost python3[43131]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/service_configs.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:29 localhost python3[43149]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/service_configs.json _original_basename=service_configs.j2 recurse=False state=file path=/etc/puppet/hieradata/service_configs.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:29 localhost python3[43211]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/extraconfig.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:30 localhost python3[43229]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/extraconfig.json _original_basename=extraconfig.j2 recurse=False state=file path=/etc/puppet/hieradata/extraconfig.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:30 localhost python3[43291]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/role_extraconfig.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:30 localhost python3[43309]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/role_extraconfig.json _original_basename=role_extraconfig.j2 recurse=False state=file path=/etc/puppet/hieradata/role_extraconfig.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:31 localhost python3[43371]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/ovn_chassis_mac_map.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:31 localhost python3[43389]: ansible-ansible.legacy.file Invoked with mode=None dest=/etc/puppet/hieradata/ovn_chassis_mac_map.json _original_basename=ovn_chassis_mac_map.j2 recurse=False state=file path=/etc/puppet/hieradata/ovn_chassis_mac_map.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:32 localhost python3[43419]: ansible-stat Invoked with path={'src': '/etc/puppet/hieradata/ansible_managed.json'} follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:56:32 localhost python3[43467]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/ansible_managed.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:33 localhost python3[43485]: ansible-ansible.legacy.file Invoked with owner=root group=root mode=0644 dest=/etc/puppet/hieradata/ansible_managed.json _original_basename=tmpyzhidlzy recurse=False state=file path=/etc/puppet/hieradata/ansible_managed.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:36 localhost python3[43515]: ansible-dnf Invoked with name=['firewalld'] state=absent allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:56:41 localhost python3[43532]: ansible-ansible.builtin.systemd Invoked with name=iptables.service state=stopped enabled=False daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:56:42 localhost python3[43550]: ansible-ansible.builtin.systemd Invoked with name=ip6tables.service state=stopped enabled=False daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:56:42 localhost python3[43568]: ansible-ansible.builtin.systemd Invoked with name=nftables state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:56:42 localhost systemd[1]: Reloading. Nov 27 02:56:42 localhost systemd-rc-local-generator[43592]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:56:42 localhost systemd-sysv-generator[43597]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:56:42 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:56:42 localhost systemd[1]: Starting Netfilter Tables... Nov 27 02:56:42 localhost systemd[1]: Finished Netfilter Tables. Nov 27 02:56:43 localhost python3[43657]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/iptables.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:43 localhost python3[43700]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/iptables.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230203.1646426-74173-102514174732418/source _original_basename=iptables.nft follow=False checksum=ede9860c99075946a7bc827210247aac639bc84a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:44 localhost python3[43730]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/iptables.nft _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:44 localhost python3[43748]: ansible-ansible.legacy.command Invoked with _raw_params=nft -j list ruleset _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:45 localhost python3[43797]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/tripleo-jumps.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:45 localhost python3[43840]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/tripleo-jumps.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230204.8935513-74280-183276022767226/source mode=None follow=False _original_basename=jump-chain.j2 checksum=eec306c3276262a27663d76bd0ea526457445afa backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:47 localhost python3[43902]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/tripleo-update-jumps.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:47 localhost python3[43945]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/tripleo-update-jumps.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230206.7894797-74353-236154514397724/source mode=None follow=False _original_basename=jump-chain.j2 checksum=eec306c3276262a27663d76bd0ea526457445afa backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:48 localhost python3[44007]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/tripleo-flushes.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:48 localhost python3[44050]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/tripleo-flushes.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230207.836168-74419-55427283233829/source mode=None follow=False _original_basename=flush-chain.j2 checksum=e8e7b8db0d61a7fe393441cc91613f470eb34a6e backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:49 localhost python3[44112]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/tripleo-chains.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:49 localhost python3[44155]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/tripleo-chains.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230208.8102324-74471-60332198102721/source mode=None follow=False _original_basename=chains.j2 checksum=e60ee651f5014e83924f4e901ecc8e25b1906610 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:50 localhost python3[44217]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/tripleo-rules.nft follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:56:50 localhost python3[44260]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/tripleo-rules.nft src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230209.7916229-74524-117075305605009/source mode=None follow=False _original_basename=ruleset.j2 checksum=0444e4206083f91e2fb2aabfa2928244c2db35ed backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:51 localhost python3[44290]: ansible-ansible.legacy.command Invoked with _raw_params=cat /etc/nftables/tripleo-chains.nft /etc/nftables/tripleo-flushes.nft /etc/nftables/tripleo-rules.nft /etc/nftables/tripleo-update-jumps.nft /etc/nftables/tripleo-jumps.nft | nft -c -f - _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:51 localhost python3[44355]: ansible-ansible.builtin.blockinfile Invoked with path=/etc/sysconfig/nftables.conf backup=False validate=nft -c -f %s block=include "/etc/nftables/iptables.nft"#012include "/etc/nftables/tripleo-chains.nft"#012include "/etc/nftables/tripleo-rules.nft"#012include "/etc/nftables/tripleo-jumps.nft"#012 state=present marker=# {mark} ANSIBLE MANAGED BLOCK create=False marker_begin=BEGIN marker_end=END unsafe_writes=False insertafter=None insertbefore=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:56:52 localhost python3[44372]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/tripleo-chains.nft _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:52 localhost python3[44389]: ansible-ansible.legacy.command Invoked with _raw_params=cat /etc/nftables/tripleo-flushes.nft /etc/nftables/tripleo-rules.nft /etc/nftables/tripleo-update-jumps.nft | nft -f - _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:56:53 localhost python3[44408]: ansible-file Invoked with mode=0750 path=/var/log/containers/collectd setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:56:53 localhost python3[44424]: ansible-file Invoked with mode=0755 path=/var/lib/container-user-scripts/ setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:56:53 localhost python3[44440]: ansible-file Invoked with mode=0750 path=/var/log/containers/ceilometer setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:56:54 localhost python3[44456]: ansible-seboolean Invoked with name=virt_sandbox_use_netlink persistent=True state=True ignore_selinux_state=False Nov 27 02:56:55 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=7 res=1 Nov 27 02:56:55 localhost python3[44476]: ansible-community.general.sefcontext Invoked with setype=container_file_t state=present target=/etc/iscsi(/.*)? ignore_selinux_state=False ftype=a reload=True seuser=None selevel=None Nov 27 02:56:56 localhost kernel: SELinux: Converting 2703 SID table entries... Nov 27 02:56:56 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:56:56 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:56:56 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:56:56 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:56:56 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:56:56 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:56:56 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:56:56 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=8 res=1 Nov 27 02:56:56 localhost python3[44497]: ansible-community.general.sefcontext Invoked with setype=container_file_t state=present target=/etc/target(/.*)? ignore_selinux_state=False ftype=a reload=True seuser=None selevel=None Nov 27 02:56:57 localhost kernel: SELinux: Converting 2703 SID table entries... Nov 27 02:56:57 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:56:57 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:56:57 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:56:57 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:56:57 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:56:57 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:56:57 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:56:57 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=9 res=1 Nov 27 02:56:58 localhost python3[44518]: ansible-community.general.sefcontext Invoked with setype=container_file_t state=present target=/var/lib/iscsi(/.*)? ignore_selinux_state=False ftype=a reload=True seuser=None selevel=None Nov 27 02:56:58 localhost kernel: SELinux: Converting 2703 SID table entries... Nov 27 02:56:58 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:56:58 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:56:58 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:56:58 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:56:58 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:56:58 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:56:58 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:56:59 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=10 res=1 Nov 27 02:56:59 localhost python3[44539]: ansible-file Invoked with path=/etc/iscsi setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:56:59 localhost python3[44555]: ansible-file Invoked with path=/etc/target setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:56:59 localhost python3[44571]: ansible-file Invoked with path=/var/lib/iscsi setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:00 localhost python3[44587]: ansible-stat Invoked with path=/lib/systemd/system/iscsid.socket follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:57:00 localhost python3[44603]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-enabled --quiet iscsi.service _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:01 localhost python3[44620]: ansible-ansible.legacy.dnf Invoked with name=['dracut-config-generic'] state=absent allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:57:05 localhost python3[44637]: ansible-file Invoked with path=/etc/modules-load.d state=directory mode=493 owner=root group=root setype=etc_t recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:05 localhost python3[44685]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/99-tripleo.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:05 localhost python3[44728]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230225.2212107-75453-169526564158424/source dest=/etc/modules-load.d/99-tripleo.conf mode=420 owner=root group=root setype=etc_t follow=False _original_basename=tripleo-modprobe.conf.j2 checksum=8021efe01721d8fa8cab46b95c00ec1be6dbb9d0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:06 localhost python3[44758]: ansible-systemd Invoked with name=systemd-modules-load.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:57:06 localhost systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 27 02:57:06 localhost systemd[1]: Stopped Load Kernel Modules. Nov 27 02:57:06 localhost systemd[1]: Stopping Load Kernel Modules... Nov 27 02:57:06 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 02:57:06 localhost kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 27 02:57:06 localhost kernel: Bridge firewalling registered Nov 27 02:57:06 localhost systemd-modules-load[44761]: Inserted module 'br_netfilter' Nov 27 02:57:06 localhost systemd-modules-load[44761]: Module 'msr' is built in Nov 27 02:57:06 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 02:57:07 localhost python3[44812]: ansible-ansible.legacy.stat Invoked with path=/etc/sysctl.d/99-tripleo.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:07 localhost python3[44855]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230226.6550434-75532-37789938702157/source dest=/etc/sysctl.d/99-tripleo.conf mode=420 owner=root group=root setype=etc_t follow=False _original_basename=tripleo-sysctl.conf.j2 checksum=cddb9401fdafaaf28a4a94b98448f98ae93c94c9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:08 localhost python3[44885]: ansible-sysctl Invoked with name=fs.aio-max-nr value=1048576 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:08 localhost python3[44902]: ansible-sysctl Invoked with name=fs.inotify.max_user_instances value=1024 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:08 localhost python3[44920]: ansible-sysctl Invoked with name=kernel.pid_max value=1048576 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:08 localhost python3[44938]: ansible-sysctl Invoked with name=net.bridge.bridge-nf-call-arptables value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:09 localhost python3[44955]: ansible-sysctl Invoked with name=net.bridge.bridge-nf-call-ip6tables value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:09 localhost python3[44972]: ansible-sysctl Invoked with name=net.bridge.bridge-nf-call-iptables value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:10 localhost python3[44989]: ansible-sysctl Invoked with name=net.ipv4.conf.all.rp_filter value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:11 localhost python3[45007]: ansible-sysctl Invoked with name=net.ipv4.ip_forward value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:11 localhost python3[45025]: ansible-sysctl Invoked with name=net.ipv4.ip_local_reserved_ports value=35357,49000-49001 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:11 localhost python3[45043]: ansible-sysctl Invoked with name=net.ipv4.ip_nonlocal_bind value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:12 localhost python3[45061]: ansible-sysctl Invoked with name=net.ipv4.neigh.default.gc_thresh1 value=1024 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:12 localhost python3[45079]: ansible-sysctl Invoked with name=net.ipv4.neigh.default.gc_thresh2 value=2048 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:12 localhost python3[45097]: ansible-sysctl Invoked with name=net.ipv4.neigh.default.gc_thresh3 value=4096 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:13 localhost python3[45115]: ansible-sysctl Invoked with name=net.ipv6.conf.all.disable_ipv6 value=0 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:13 localhost python3[45132]: ansible-sysctl Invoked with name=net.ipv6.conf.all.forwarding value=0 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:13 localhost python3[45149]: ansible-sysctl Invoked with name=net.ipv6.conf.default.disable_ipv6 value=0 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:13 localhost python3[45166]: ansible-sysctl Invoked with name=net.ipv6.conf.lo.disable_ipv6 value=0 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:14 localhost python3[45183]: ansible-sysctl Invoked with name=net.ipv6.ip_nonlocal_bind value=1 sysctl_set=True state=present sysctl_file=/etc/sysctl.d/99-tripleo.conf reload=False ignoreerrors=False Nov 27 02:57:14 localhost python3[45201]: ansible-systemd Invoked with name=systemd-sysctl.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 02:57:14 localhost systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 27 02:57:14 localhost systemd[1]: Stopped Apply Kernel Variables. Nov 27 02:57:14 localhost systemd[1]: Stopping Apply Kernel Variables... Nov 27 02:57:14 localhost systemd[1]: Starting Apply Kernel Variables... Nov 27 02:57:14 localhost systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 27 02:57:14 localhost systemd[1]: Finished Apply Kernel Variables. Nov 27 02:57:15 localhost python3[45221]: ansible-file Invoked with mode=0750 path=/var/log/containers/metrics_qdr setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:15 localhost python3[45237]: ansible-file Invoked with path=/var/lib/metrics_qdr setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:15 localhost python3[45253]: ansible-file Invoked with mode=0750 path=/var/log/containers/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:16 localhost python3[45269]: ansible-stat Invoked with path=/var/lib/nova/instances follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:57:16 localhost python3[45285]: ansible-file Invoked with path=/var/lib/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:16 localhost python3[45301]: ansible-file Invoked with path=/var/lib/_nova_secontext setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:17 localhost python3[45317]: ansible-file Invoked with path=/var/lib/nova/instances setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:17 localhost python3[45333]: ansible-file Invoked with path=/var/lib/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:17 localhost python3[45349]: ansible-file Invoked with path=/etc/tmpfiles.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:18 localhost python3[45397]: ansible-ansible.legacy.stat Invoked with path=/etc/tmpfiles.d/run-nova.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:18 localhost python3[45440]: ansible-ansible.legacy.copy Invoked with dest=/etc/tmpfiles.d/run-nova.conf src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230237.942281-75898-241532057847515/source _original_basename=tmply_latbi follow=False checksum=f834349098718ec09c7562bcb470b717a83ff411 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:19 localhost python3[45470]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-tmpfiles --create _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:20 localhost python3[45487]: ansible-file Invoked with path=/var/lib/tripleo-config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:20 localhost python3[45535]: ansible-ansible.legacy.stat Invoked with path=/var/lib/nova/delay-nova-compute follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:21 localhost python3[45578]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/nova/delay-nova-compute mode=493 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230240.4063287-76025-4335577952908/source _original_basename=tmpr_ya1ywh follow=False checksum=f07ad3e8cf3766b3b3b07ae8278826a0ef3bb5e3 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:21 localhost python3[45608]: ansible-file Invoked with mode=0750 path=/var/log/containers/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:21 localhost python3[45624]: ansible-file Invoked with path=/etc/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:22 localhost python3[45640]: ansible-file Invoked with path=/etc/libvirt/secrets setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:22 localhost python3[45656]: ansible-file Invoked with path=/etc/libvirt/qemu setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:22 localhost python3[45672]: ansible-file Invoked with path=/var/lib/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:23 localhost python3[45688]: ansible-file Invoked with path=/var/cache/libvirt state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:23 localhost python3[45704]: ansible-file Invoked with path=/var/lib/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:23 localhost python3[45720]: ansible-file Invoked with path=/run/libvirt state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:24 localhost python3[45736]: ansible-file Invoked with mode=0770 path=/var/log/containers/libvirt/swtpm setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:24 localhost python3[45752]: ansible-group Invoked with gid=107 name=qemu state=present system=False local=False non_unique=False Nov 27 02:57:24 localhost python3[45774]: ansible-user Invoked with comment=qemu user group=qemu name=qemu shell=/sbin/nologin state=present uid=107 non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.localdomain update_password=always groups=None home=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Nov 27 02:57:25 localhost python3[45798]: ansible-file Invoked with group=qemu owner=qemu path=/var/lib/vhost_sockets setype=virt_cache_t seuser=system_u state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None serole=None selevel=None attributes=None Nov 27 02:57:25 localhost python3[45814]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/bin/rpm -q libvirt-daemon _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:26 localhost python3[45863]: ansible-ansible.legacy.stat Invoked with path=/etc/tmpfiles.d/run-libvirt.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:26 localhost python3[45906]: ansible-ansible.legacy.copy Invoked with dest=/etc/tmpfiles.d/run-libvirt.conf src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230245.7563503-76330-85065013634483/source _original_basename=tmpzje_0lwe follow=False checksum=57f3ff94c666c6aae69ae22e23feb750cf9e8b13 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:26 localhost python3[45936]: ansible-seboolean Invoked with name=os_enable_vtpm persistent=True state=True ignore_selinux_state=False Nov 27 02:57:27 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=11 res=1 Nov 27 02:57:27 localhost python3[46021]: ansible-file Invoked with path=/etc/crypto-policies/local.d/gnutls-qemu.config state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:28 localhost python3[46067]: ansible-file Invoked with path=/run/libvirt setype=virt_var_run_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:28 localhost python3[46100]: ansible-seboolean Invoked with name=logrotate_read_inside_containers persistent=True state=True ignore_selinux_state=False Nov 27 02:57:28 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=12 res=1 Nov 27 02:57:30 localhost python3[46135]: ansible-ansible.legacy.dnf Invoked with name=['podman'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:57:33 localhost python3[46152]: ansible-setup Invoked with gather_subset=['!all', '!min', 'network'] filter=['ansible_interfaces'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 02:57:34 localhost python3[46213]: ansible-file Invoked with path=/etc/containers/networks state=directory recurse=True mode=493 owner=root group=root force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:34 localhost python3[46229]: ansible-ansible.legacy.command Invoked with _raw_params=podman network inspect podman#012 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:34 localhost python3[46289]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/networks/podman.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:35 localhost python3[46332]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230254.5653303-76869-240769974692646/source dest=/etc/containers/networks/podman.json mode=0644 owner=root group=root follow=False _original_basename=podman_network_config.j2 checksum=e5aa92f4abcfd96b677e09d067326c1ed269e204 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:35 localhost python3[46394]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:36 localhost python3[46439]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230255.6086786-76914-251431975973613/source dest=/etc/containers/registries.conf owner=root group=root setype=etc_t mode=0644 follow=False _original_basename=registries.conf.j2 checksum=710a00cfb11a4c3eba9c028ef1984a9fea9ba83a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:36 localhost python3[46469]: ansible-ini_file Invoked with path=/etc/containers/containers.conf owner=root group=root setype=etc_t mode=0644 create=True section=containers option=pids_limit value=4096 backup=False state=present exclusive=True no_extra_spaces=False allow_no_value=False unsafe_writes=False values=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:37 localhost python3[46485]: ansible-ini_file Invoked with path=/etc/containers/containers.conf owner=root group=root setype=etc_t mode=0644 create=True section=engine option=events_logger value="journald" backup=False state=present exclusive=True no_extra_spaces=False allow_no_value=False unsafe_writes=False values=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:37 localhost python3[46501]: ansible-ini_file Invoked with path=/etc/containers/containers.conf owner=root group=root setype=etc_t mode=0644 create=True section=engine option=runtime value="crun" backup=False state=present exclusive=True no_extra_spaces=False allow_no_value=False unsafe_writes=False values=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:37 localhost python3[46517]: ansible-ini_file Invoked with path=/etc/containers/containers.conf owner=root group=root setype=etc_t mode=0644 create=True section=network option=network_backend value="netavark" backup=False state=present exclusive=True no_extra_spaces=False allow_no_value=False unsafe_writes=False values=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:38 localhost python3[46565]: ansible-ansible.legacy.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:38 localhost python3[46608]: ansible-ansible.legacy.copy Invoked with dest=/etc/sysconfig/podman_drop_in src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230258.1683054-77069-244952156647972/source _original_basename=tmpmlvvvxub follow=False checksum=0bfbc70e9a4740c9004b9947da681f723d529c83 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:39 localhost python3[46638]: ansible-file Invoked with mode=0750 path=/var/log/containers/rsyslog setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:39 localhost python3[46654]: ansible-file Invoked with path=/var/lib/rsyslog.container setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:40 localhost python3[46670]: ansible-ansible.legacy.dnf Invoked with name=['openssh-server'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:57:43 localhost python3[46719]: ansible-ansible.legacy.stat Invoked with path=/etc/ssh/sshd_config follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:44 localhost python3[46764]: ansible-ansible.legacy.copy Invoked with dest=/etc/ssh/sshd_config src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230263.5705342-77372-238237661130344/source validate=/usr/sbin/sshd -T -f %s mode=None follow=False _original_basename=sshd_config_block.j2 checksum=913c99ed7d5c33615bfb07a6792a4ef143dcfd2b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:44 localhost python3[46795]: ansible-systemd Invoked with name=sshd state=restarted enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:57:44 localhost systemd[1]: Stopping OpenSSH server daemon... Nov 27 02:57:44 localhost systemd[1]: sshd.service: Deactivated successfully. Nov 27 02:57:44 localhost systemd[1]: Stopped OpenSSH server daemon. Nov 27 02:57:44 localhost systemd[1]: sshd.service: Consumed 1.937s CPU time, read 1.9M from disk, written 0B to disk. Nov 27 02:57:44 localhost systemd[1]: Stopped target sshd-keygen.target. Nov 27 02:57:44 localhost systemd[1]: Stopping sshd-keygen.target... Nov 27 02:57:44 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 02:57:44 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 02:57:44 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 02:57:44 localhost systemd[1]: Reached target sshd-keygen.target. Nov 27 02:57:44 localhost systemd[1]: Starting OpenSSH server daemon... Nov 27 02:57:44 localhost sshd[46799]: main: sshd: ssh-rsa algorithm is disabled Nov 27 02:57:44 localhost systemd[1]: Started OpenSSH server daemon. Nov 27 02:57:45 localhost python3[46815]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-active ntpd.service || systemctl is-enabled ntpd.service _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:46 localhost python3[46833]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-active ntpd.service || systemctl is-enabled ntpd.service _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:46 localhost python3[46851]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:57:50 localhost python3[46900]: ansible-ansible.legacy.stat Invoked with path=/etc/chrony.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:50 localhost python3[46918]: ansible-ansible.legacy.file Invoked with owner=root group=root mode=420 dest=/etc/chrony.conf _original_basename=chrony.conf.j2 recurse=False state=file path=/etc/chrony.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:51 localhost python3[46948]: ansible-ansible.legacy.systemd Invoked with enabled=True name=chronyd state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 02:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 600.1 total, 600.0 interval#012Cumulative writes: 3249 writes, 16K keys, 3249 commit groups, 1.0 writes per commit group, ingest: 0.01 GB, 0.02 MB/s#012Cumulative WAL: 3249 writes, 140 syncs, 23.21 writes per sync, written: 0.01 GB, 0.02 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 3249 writes, 16K keys, 3249 commit groups, 1.0 writes per commit group, ingest: 14.61 MB, 0.02 MB/s#012Interval WAL: 3249 writes, 140 syncs, 23.21 writes per sync, written: 0.01 GB, 0.02 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 8 last_secs: 3.3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 8 last_secs: 3.3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memt Nov 27 02:57:51 localhost python3[46998]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/chrony-online.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:52 localhost python3[47016]: ansible-ansible.legacy.file Invoked with dest=/etc/systemd/system/chrony-online.service _original_basename=chrony-online.service recurse=False state=file path=/etc/systemd/system/chrony-online.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:57:52 localhost python3[47046]: ansible-systemd Invoked with state=started name=chrony-online.service enabled=True daemon-reload=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 02:57:52 localhost systemd[1]: Reloading. Nov 27 02:57:52 localhost systemd-rc-local-generator[47070]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:57:52 localhost systemd-sysv-generator[47076]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:57:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:57:53 localhost systemd[1]: Starting chronyd online sources service... Nov 27 02:57:53 localhost chronyc[47086]: 200 OK Nov 27 02:57:53 localhost systemd[1]: chrony-online.service: Deactivated successfully. Nov 27 02:57:53 localhost systemd[1]: Finished chronyd online sources service. Nov 27 02:57:53 localhost python3[47102]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc makestep _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:53 localhost chronyd[25955]: System clock was stepped by 0.000044 seconds Nov 27 02:57:53 localhost python3[47119]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 30 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:54 localhost python3[47136]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc makestep _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:54 localhost chronyd[25955]: System clock was stepped by 0.000000 seconds Nov 27 02:57:54 localhost python3[47153]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 30 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:55 localhost python3[47170]: ansible-timezone Invoked with name=UTC hwclock=None Nov 27 02:57:55 localhost systemd[1]: Starting Time & Date Service... Nov 27 02:57:55 localhost systemd[1]: Started Time & Date Service. Nov 27 02:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 02:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 600.2 total, 600.0 interval#012Cumulative writes: 3384 writes, 16K keys, 3384 commit groups, 1.0 writes per commit group, ingest: 0.01 GB, 0.03 MB/s#012Cumulative WAL: 3384 writes, 195 syncs, 17.35 writes per sync, written: 0.01 GB, 0.03 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 3384 writes, 16K keys, 3384 commit groups, 1.0 writes per commit group, ingest: 15.24 MB, 0.03 MB/s#012Interval WAL: 3384 writes, 195 syncs, 17.35 writes per sync, written: 0.01 GB, 0.03 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 8 last_secs: 5.7e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 8 last_secs: 5.7e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memt Nov 27 02:57:56 localhost python3[47190]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q --whatprovides tuned tuned-profiles-cpu-partitioning _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:56 localhost python3[47207]: ansible-ansible.legacy.command Invoked with _raw_params=which tuned-adm _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:57:57 localhost python3[47224]: ansible-slurp Invoked with src=/etc/tuned/active_profile Nov 27 02:57:57 localhost python3[47240]: ansible-stat Invoked with path=/etc/tuned/throughput-performance-variables.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:57:58 localhost python3[47256]: ansible-file Invoked with mode=0750 path=/var/log/containers/openvswitch setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:58 localhost python3[47272]: ansible-file Invoked with path=/var/lib/openvswitch/ovn setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:57:59 localhost python3[47320]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/neutron-cleanup follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:57:59 localhost python3[47363]: ansible-ansible.legacy.copy Invoked with dest=/usr/libexec/neutron-cleanup force=True mode=0755 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230278.948208-78275-129502713572078/source _original_basename=tmpg4g2nr_i follow=False checksum=f9cc7d1e91fbae49caa7e35eb2253bba146a73b4 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:00 localhost python3[47425]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/neutron-cleanup.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:58:00 localhost python3[47468]: ansible-ansible.legacy.copy Invoked with dest=/usr/lib/systemd/system/neutron-cleanup.service force=True src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230279.8787866-78337-60238743459034/source _original_basename=tmpwwcbbmg0 follow=False checksum=6b6cd9f074903a28d054eb530a10c7235d0c39fc backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:01 localhost python3[47498]: ansible-ansible.legacy.systemd Invoked with enabled=True name=neutron-cleanup daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 02:58:01 localhost systemd[1]: Reloading. Nov 27 02:58:01 localhost systemd-rc-local-generator[47528]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:58:01 localhost systemd-sysv-generator[47531]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:58:01 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:58:01 localhost python3[47552]: ansible-file Invoked with mode=0750 path=/var/log/containers/neutron setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:58:02 localhost python3[47568]: ansible-ansible.legacy.command Invoked with _raw_params=ip netns add ns_temp _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:58:02 localhost python3[47585]: ansible-ansible.legacy.command Invoked with _raw_params=ip netns delete ns_temp _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:58:02 localhost systemd[1]: run-netns-ns_temp.mount: Deactivated successfully. Nov 27 02:58:02 localhost python3[47602]: ansible-file Invoked with path=/var/lib/neutron setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:58:03 localhost python3[47618]: ansible-file Invoked with path=/var/lib/neutron/kill_scripts state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:03 localhost python3[47666]: ansible-ansible.legacy.stat Invoked with path=/var/lib/neutron/kill_scripts/haproxy-kill follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:58:03 localhost python3[47709]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/neutron/kill_scripts/haproxy-kill mode=493 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230283.259539-78563-105235459158805/source _original_basename=tmpq07t6zqs follow=False checksum=2f369fbe8f83639cdfd4efc53e7feb4ee77d1ed7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:25 localhost systemd[1]: systemd-timedated.service: Deactivated successfully. Nov 27 02:58:30 localhost python3[47802]: ansible-file Invoked with path=/var/log/containers state=directory setype=container_file_t selevel=s0 mode=488 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 02:58:30 localhost python3[47818]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory selevel=s0 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None setype=None attributes=None Nov 27 02:58:30 localhost python3[47834]: ansible-file Invoked with path=/var/lib/tripleo-config state=directory setype=container_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 02:58:31 localhost python3[47850]: ansible-file Invoked with path=/var/lib/container-startup-configs.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:31 localhost python3[47866]: ansible-file Invoked with path=/var/lib/docker-container-startup-configs.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:32 localhost python3[47897]: ansible-community.general.sefcontext Invoked with target=/var/lib/container-config-scripts(/.*)? setype=container_file_t state=present ignore_selinux_state=False ftype=a reload=True seuser=None selevel=None Nov 27 02:58:32 localhost kernel: SELinux: Converting 2706 SID table entries... Nov 27 02:58:32 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:58:32 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:58:32 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:58:32 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:58:32 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:58:32 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:58:32 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:58:33 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=13 res=1 Nov 27 02:58:33 localhost python3[47918]: ansible-file Invoked with path=/var/lib/container-config-scripts state=directory setype=container_file_t recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 02:58:34 localhost systemd[36158]: Created slice User Background Tasks Slice. Nov 27 02:58:34 localhost systemd[36158]: Starting Cleanup of User's Temporary Files and Directories... Nov 27 02:58:34 localhost systemd[36158]: Finished Cleanup of User's Temporary Files and Directories. Nov 27 02:58:35 localhost python3[48056]: ansible-container_startup_config Invoked with config_base_dir=/var/lib/tripleo-config/container-startup-config config_data={'step_1': {'metrics_qdr': {'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, 'metrics_qdr_init_logs': {'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}}, 'step_2': {'create_haproxy_wrapper': {'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, 'create_virtlogd_wrapper': {'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, 'nova_compute_init_log': {'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, 'nova_virtqemud_init_logs': {'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}}, 'step_3': {'ceilometer_init_log': {'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, 'collectd': {'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, 'iscsid': {'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, 'nova_statedir_owner': {'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, 'nova_virtlogd_wrapper': {'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': [ Nov 27 02:58:35 localhost rsyslogd[760]: message too long (31243) with configured size 8096, begin of message is: ansible-container_startup_config Invoked with config_base_dir=/var/lib/tripleo-c [v8.2102.0-111.el9 try https://www.rsyslog.com/e/2445 ] Nov 27 02:58:36 localhost python3[48072]: ansible-file Invoked with path=/var/lib/kolla/config_files state=directory setype=container_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 02:58:36 localhost python3[48088]: ansible-file Invoked with path=/var/lib/config-data mode=493 state=directory setype=container_file_t selevel=s0 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 02:58:36 localhost python3[48104]: ansible-tripleo_container_configs Invoked with config_data={'/var/lib/kolla/config_files/ceilometer-agent-ipmi.json': {'command': '/usr/bin/ceilometer-polling --polling-namespaces ipmi --logfile /var/log/ceilometer/ipmi.log', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}]}, '/var/lib/kolla/config_files/ceilometer_agent_compute.json': {'command': '/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /var/log/ceilometer/compute.log', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}]}, '/var/lib/kolla/config_files/collectd.json': {'command': '/usr/sbin/collectd -f', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/', 'merge': False, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/etc/collectd.d'}], 'permissions': [{'owner': 'collectd:collectd', 'path': '/var/log/collectd', 'recurse': True}, {'owner': 'collectd:collectd', 'path': '/scripts', 'recurse': True}, {'owner': 'collectd:collectd', 'path': '/config-scripts', 'recurse': True}]}, '/var/lib/kolla/config_files/iscsid.json': {'command': '/usr/sbin/iscsid -f', 'config_files': [{'dest': '/etc/iscsi/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-iscsid/'}]}, '/var/lib/kolla/config_files/logrotate-crond.json': {'command': '/usr/sbin/crond -s -n', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}]}, '/var/lib/kolla/config_files/metrics_qdr.json': {'command': '/usr/sbin/qdrouterd -c /etc/qpid-dispatch/qdrouterd.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/', 'merge': True, 'optional': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-tls/*'}], 'permissions': [{'owner': 'qdrouterd:qdrouterd', 'path': '/var/lib/qdrouterd', 'recurse': True}, {'optional': True, 'owner': 'qdrouterd:qdrouterd', 'path': '/etc/pki/tls/certs/metrics_qdr.crt'}, {'optional': True, 'owner': 'qdrouterd:qdrouterd', 'path': '/etc/pki/tls/private/metrics_qdr.key'}]}, '/var/lib/kolla/config_files/nova-migration-target.json': {'command': 'dumb-init --single-child -- /usr/sbin/sshd -D -p 2022', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ssh/', 'owner': 'root', 'perm': '0600', 'source': '/host-ssh/ssh_host_*_key'}]}, '/var/lib/kolla/config_files/nova_compute.json': {'command': '/var/lib/nova/delay-nova-compute --delay 180 --nova-binary /usr/bin/nova-compute ', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/iscsi/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-iscsid/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/var/log/nova', 'recurse': True}, {'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json': {'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_wait_for_compute_service.py', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}], 'permissions': [{'owner': 'nova:nova', 'path': '/var/log/nova', 'recurse': True}]}, '/var/lib/kolla/config_files/nova_virtlogd.json': {'command': '/usr/local/bin/virtlogd_wrapper', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_virtnodedevd.json': {'command': '/usr/sbin/virtnodedevd --config /etc/libvirt/virtnodedevd.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_virtproxyd.json': {'command': '/usr/sbin/virtproxyd --config /etc/libvirt/virtproxyd.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_virtqemud.json': {'command': '/usr/sbin/virtqemud --config /etc/libvirt/virtqemud.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_virtsecretd.json': {'command': '/usr/sbin/virtsecretd --config /etc/libvirt/virtsecretd.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/nova_virtstoraged.json': {'command': '/usr/sbin/virtstoraged --config /etc/libvirt/virtstoraged.conf', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}, {'dest': '/etc/ceph/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src-ceph/'}], 'permissions': [{'owner': 'nova:nova', 'path': '/etc/ceph/ceph.client.openstack.keyring', 'perm': '0600'}]}, '/var/lib/kolla/config_files/ovn_controller.json': {'command': '/usr/bin/ovn-controller --pidfile --log-file unix:/run/openvswitch/db.sock ', 'permissions': [{'owner': 'root:root', 'path': '/var/log/openvswitch', 'recurse': True}, {'owner': 'root:root', 'path': '/var/log/ovn', 'recurse': True}]}, '/var/lib/kolla/config_files/ovn_metadata_agent.json': {'command': '/usr/bin/networking-ovn-metadata-agent --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/networking-ovn/networking-ovn-metadata-agent.ini --log-file=/var/log/neutron/ovn-metadata-agent.log', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}], 'permissions': [{'owner': 'neutron:neutron', 'path': '/var/log/neutron', 'recurse': True}, {'owner': 'neutron:neutron', 'path': '/var/lib/neutron', 'recurse': True}, {'optional': True, 'owner': 'neutron:neutron', 'path': '/etc/pki/tls/certs/ovn_metadata.crt', 'perm': '0644'}, {'optional': True, 'owner': 'neutron:neutron', 'path': '/etc/pki/tls/private/ovn_metadata.key', 'perm': '0644'}]}, '/var/lib/kolla/config_files/rsyslog.json': {'command': '/usr/sbin/rsyslogd -n', 'config_files': [{'dest': '/', 'merge': True, 'preserve_properties': True, 'source': '/var/lib/kolla/config_files/src/*'}], 'permissions': [{'owner': 'root:root', 'path': '/var/lib/rsyslog', 'recurse': True}, {'owner': 'root:root', 'path': '/var/log/rsyslog', 'recurse': True}]}} Nov 27 02:58:42 localhost python3[48152]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/config_step.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 02:58:42 localhost python3[48195]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/config_step.json force=True mode=0600 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230322.251716-80305-39653408189358/source _original_basename=tmpfb679eyq follow=False checksum=dfdcc7695edd230e7a2c06fc7b739bfa56506d8f backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 02:58:43 localhost python3[48225]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_1 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 02:58:46 localhost python3[48348]: ansible-file Invoked with path=/var/lib/container-puppet state=directory setype=container_file_t selevel=s0 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 02:58:48 localhost python3[48469]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 02:58:50 localhost python3[48485]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -q lvm2 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:58:51 localhost python3[48502]: ansible-ansible.legacy.dnf Invoked with name=['systemd-container'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 02:58:56 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:58:56 localhost dbus-broker-launch[18428]: Noticed file-system modification, trigger reload. Nov 27 02:58:56 localhost dbus-broker-launch[18428]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +31: Eavesdropping is deprecated and ignored Nov 27 02:58:56 localhost dbus-broker-launch[18428]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +33: Eavesdropping is deprecated and ignored Nov 27 02:58:56 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:58:56 localhost systemd[1]: Reexecuting. Nov 27 02:58:57 localhost systemd[1]: systemd 252-14.el9_2.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN -IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK +XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Nov 27 02:58:57 localhost systemd[1]: Detected virtualization kvm. Nov 27 02:58:57 localhost systemd[1]: Detected architecture x86-64. Nov 27 02:58:57 localhost systemd-rc-local-generator[48556]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:58:57 localhost systemd-sysv-generator[48559]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:58:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:59:05 localhost kernel: SELinux: Converting 2706 SID table entries... Nov 27 02:59:05 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 02:59:05 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 02:59:05 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 02:59:05 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 02:59:05 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 02:59:05 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 02:59:05 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 02:59:05 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:59:05 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=14 res=1 Nov 27 02:59:05 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 02:59:06 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:59:06 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 02:59:06 localhost systemd[1]: Reloading. Nov 27 02:59:06 localhost systemd-sysv-generator[48673]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:59:06 localhost systemd-rc-local-generator[48668]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:59:06 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:59:06 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 02:59:06 localhost systemd[1]: Stopping Journal Service... Nov 27 02:59:06 localhost systemd[1]: Stopping Rule-based Manager for Device Events and Files... Nov 27 02:59:06 localhost systemd-journald[618]: Received SIGTERM from PID 1 (systemd). Nov 27 02:59:06 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:59:06 localhost systemd-journald[618]: Journal stopped Nov 27 02:59:06 localhost systemd[1]: systemd-journald.service: Deactivated successfully. Nov 27 02:59:06 localhost systemd[1]: Stopped Journal Service. Nov 27 02:59:06 localhost systemd[1]: systemd-journald.service: Consumed 1.706s CPU time. Nov 27 02:59:06 localhost systemd[1]: Starting Journal Service... Nov 27 02:59:06 localhost systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 27 02:59:06 localhost systemd[1]: Stopped Rule-based Manager for Device Events and Files. Nov 27 02:59:06 localhost systemd[1]: systemd-udevd.service: Consumed 3.162s CPU time. Nov 27 02:59:06 localhost systemd[1]: Starting Rule-based Manager for Device Events and Files... Nov 27 02:59:06 localhost systemd-journald[49007]: Journal started Nov 27 02:59:06 localhost systemd-journald[49007]: Runtime Journal (/run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a) is 12.1M, max 314.7M, 302.6M free. Nov 27 02:59:06 localhost systemd[1]: Started Journal Service. Nov 27 02:59:06 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.4 (251 of 333 items), suggesting rotation. Nov 27 02:59:06 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 02:59:06 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:59:06 localhost systemd-udevd[49012]: Using default interface naming scheme 'rhel-9.0'. Nov 27 02:59:06 localhost systemd[1]: Started Rule-based Manager for Device Events and Files. Nov 27 02:59:06 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 02:59:07 localhost systemd[1]: Reloading. Nov 27 02:59:07 localhost systemd-rc-local-generator[49585]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 02:59:07 localhost systemd-sysv-generator[49593]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 02:59:07 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 02:59:07 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 02:59:07 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 02:59:07 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 02:59:07 localhost systemd[1]: man-db-cache-update.service: Consumed 1.439s CPU time. Nov 27 02:59:07 localhost systemd[1]: run-re43977522ba64266929acddfb754c259.service: Deactivated successfully. Nov 27 02:59:07 localhost systemd[1]: run-rcdf4c2d45427415186a4a505358d6490.service: Deactivated successfully. Nov 27 02:59:09 localhost python3[49996]: ansible-sysctl Invoked with name=vm.unprivileged_userfaultfd reload=True state=present sysctl_file=/etc/sysctl.d/99-tripleo-postcopy.conf sysctl_set=True value=1 ignoreerrors=False Nov 27 02:59:09 localhost python3[50015]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-active ksm.service || systemctl is-enabled ksm.service _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 02:59:10 localhost python3[50033]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 02:59:10 localhost python3[50033]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 --format json Nov 27 02:59:10 localhost python3[50033]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 -q --tls-verify=false Nov 27 02:59:17 localhost podman[50045]: 2025-11-27 07:59:11.030485673 +0000 UTC m=+0.042989785 image pull registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 02:59:18 localhost python3[50033]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect bac901955dcf7a32a493c6ef724c092009bbc18467858aa8c55e916b8c2b2b8f --format json Nov 27 02:59:18 localhost python3[50149]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 02:59:18 localhost python3[50149]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 --format json Nov 27 02:59:18 localhost python3[50149]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 -q --tls-verify=false Nov 27 02:59:26 localhost podman[50163]: 2025-11-27 07:59:18.514848604 +0000 UTC m=+0.044504394 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 02:59:26 localhost python3[50149]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 44feaf8d87c1d40487578230316b622680576d805efdb45dfeea6aad464b41f1 --format json Nov 27 02:59:27 localhost python3[50265]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 02:59:27 localhost python3[50265]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 --format json Nov 27 02:59:27 localhost python3[50265]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 -q --tls-verify=false Nov 27 02:59:36 localhost podman[50915]: 2025-11-27 07:59:36.03415131 +0000 UTC m=+0.062138142 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, RELEASE=main, release=553, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, architecture=x86_64, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , GIT_CLEAN=True, io.openshift.expose-services=, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 02:59:36 localhost podman[50915]: 2025-11-27 07:59:36.12725168 +0000 UTC m=+0.155238532 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, release=553, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., ceph=True, CEPH_POINT_RELEASE=, name=rhceph, com.redhat.component=rhceph-container) Nov 27 02:59:42 localhost podman[50278]: 2025-11-27 07:59:27.263155738 +0000 UTC m=+0.033040232 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 02:59:42 localhost python3[50265]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 3a088c12511c977065fdc5f1594cba7b1a79f163578a6ffd0ac4a475b8e67938 --format json Nov 27 02:59:43 localhost python3[51603]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 02:59:43 localhost python3[51603]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 --format json Nov 27 02:59:43 localhost python3[51603]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 -q --tls-verify=false Nov 27 02:59:55 localhost podman[51644]: 2025-11-27 07:59:43.174895526 +0000 UTC m=+0.037294380 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 02:59:55 localhost python3[51603]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 514d439186251360cf734cbc6d4a44c834664891872edf3798a653dfaacf10c0 --format json Nov 27 02:59:55 localhost python3[51807]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 02:59:55 localhost python3[51807]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 --format json Nov 27 02:59:55 localhost python3[51807]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 -q --tls-verify=false Nov 27 02:59:59 localhost podman[51819]: 2025-11-27 07:59:55.892966274 +0000 UTC m=+0.040248115 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 Nov 27 02:59:59 localhost python3[51807]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect a9dd7a2ac6f35cb086249f87f74e2f8e74e7e2ad5141ce2228263be6faedce26 --format json Nov 27 03:00:00 localhost python3[51909]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:00 localhost python3[51909]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 --format json Nov 27 03:00:00 localhost python3[51909]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 -q --tls-verify=false Nov 27 03:00:09 localhost podman[51921]: 2025-11-27 08:00:00.518561271 +0000 UTC m=+0.044735619 image pull registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 Nov 27 03:00:09 localhost python3[51909]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 24976907b2c2553304119aba5731a800204d664feed24ca9eb7f2b4c7d81016b --format json Nov 27 03:00:09 localhost python3[51999]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:09 localhost python3[51999]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 --format json Nov 27 03:00:09 localhost python3[51999]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 -q --tls-verify=false Nov 27 03:00:17 localhost podman[52011]: 2025-11-27 08:00:09.892433097 +0000 UTC m=+0.045582242 image pull registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 Nov 27 03:00:17 localhost python3[51999]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 57163a7b21fdbb804a27897cb6e6052a5e5c7a339c45d663e80b52375a760dcf --format json Nov 27 03:00:17 localhost python3[52087]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:17 localhost python3[52087]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 --format json Nov 27 03:00:17 localhost python3[52087]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 -q --tls-verify=false Nov 27 03:00:20 localhost podman[52100]: 2025-11-27 08:00:18.048329723 +0000 UTC m=+0.040834490 image pull registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 Nov 27 03:00:20 localhost python3[52087]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 076d82a27d63c8328729ed27ceb4291585ae18d017befe6fe353df7aa11715ae --format json Nov 27 03:00:20 localhost python3[52176]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:20 localhost python3[52176]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 --format json Nov 27 03:00:20 localhost python3[52176]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 -q --tls-verify=false Nov 27 03:00:29 localhost podman[52189]: 2025-11-27 08:00:20.817429707 +0000 UTC m=+0.043554071 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 Nov 27 03:00:29 localhost python3[52176]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect d0dbcb95546840a8d088df044347a7877ad5ea45a2ddba0578e9bb5de4ab0da5 --format json Nov 27 03:00:29 localhost python3[52268]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:29 localhost python3[52268]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 --format json Nov 27 03:00:29 localhost python3[52268]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 -q --tls-verify=false Nov 27 03:00:35 localhost podman[52281]: 2025-11-27 08:00:29.740388169 +0000 UTC m=+0.046775554 image pull registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:00:35 localhost python3[52268]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect e6e981540e553415b2d6eda490d7683db07164af2e7a0af8245623900338a4d6 --format json Nov 27 03:00:35 localhost python3[52387]: ansible-containers.podman.podman_image Invoked with force=True name=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 validate_certs=False tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'volume': None, 'extra_args': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} path=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Nov 27 03:00:35 localhost python3[52387]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman image ls registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 --format json Nov 27 03:00:36 localhost python3[52387]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman pull registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 -q --tls-verify=false Nov 27 03:00:38 localhost podman[52398]: 2025-11-27 08:00:36.117053044 +0000 UTC m=+0.040413686 image pull registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 Nov 27 03:00:38 localhost python3[52387]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /bin/podman inspect 87ee88cbf01fb42e0b22747072843bcca6130a90eda4de6e74b3ccd847bb4040 --format json Nov 27 03:00:38 localhost python3[52474]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_1 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:00:40 localhost ansible-async_wrapper.py[52646]: Invoked with 603233236296 3600 /home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230440.1633976-83260-10328356804471/AnsiballZ_command.py _ Nov 27 03:00:40 localhost ansible-async_wrapper.py[52649]: Starting module and watcher Nov 27 03:00:40 localhost ansible-async_wrapper.py[52649]: Start watching 52650 (3600) Nov 27 03:00:40 localhost ansible-async_wrapper.py[52650]: Start module (52650) Nov 27 03:00:40 localhost ansible-async_wrapper.py[52646]: Return async_wrapper task started. Nov 27 03:00:41 localhost python3[52670]: ansible-ansible.legacy.async_status Invoked with jid=603233236296.52646 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:00:45 localhost puppet-user[52669]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:00:45 localhost puppet-user[52669]: (file: /etc/puppet/hiera.yaml) Nov 27 03:00:45 localhost puppet-user[52669]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:00:45 localhost puppet-user[52669]: (file & line not available) Nov 27 03:00:45 localhost puppet-user[52669]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:00:45 localhost puppet-user[52669]: (file & line not available) Nov 27 03:00:45 localhost puppet-user[52669]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/database/mysql/client.pp, line: 89, column: 8) Nov 27 03:00:45 localhost puppet-user[52669]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/packages.pp, line: 39, column: 69) Nov 27 03:00:45 localhost puppet-user[52669]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.14 seconds Nov 27 03:00:45 localhost puppet-user[52669]: Notice: /Stage[main]/Tripleo::Profile::Base::Database::Mysql::Client/Exec[directory-create-etc-my.cnf.d]/returns: executed successfully Nov 27 03:00:45 localhost puppet-user[52669]: Notice: /Stage[main]/Tripleo::Profile::Base::Database::Mysql::Client/File[/etc/my.cnf.d/tripleo.cnf]/ensure: created Nov 27 03:00:45 localhost puppet-user[52669]: Notice: /Stage[main]/Tripleo::Profile::Base::Database::Mysql::Client/Augeas[tripleo-mysql-client-conf]/returns: executed successfully Nov 27 03:00:45 localhost puppet-user[52669]: Notice: Applied catalog in 0.06 seconds Nov 27 03:00:45 localhost puppet-user[52669]: Application: Nov 27 03:00:45 localhost puppet-user[52669]: Initial environment: production Nov 27 03:00:45 localhost puppet-user[52669]: Converged environment: production Nov 27 03:00:45 localhost puppet-user[52669]: Run mode: user Nov 27 03:00:45 localhost puppet-user[52669]: Changes: Nov 27 03:00:45 localhost puppet-user[52669]: Total: 3 Nov 27 03:00:45 localhost puppet-user[52669]: Events: Nov 27 03:00:45 localhost puppet-user[52669]: Success: 3 Nov 27 03:00:45 localhost puppet-user[52669]: Total: 3 Nov 27 03:00:45 localhost puppet-user[52669]: Resources: Nov 27 03:00:45 localhost puppet-user[52669]: Changed: 3 Nov 27 03:00:45 localhost puppet-user[52669]: Out of sync: 3 Nov 27 03:00:45 localhost puppet-user[52669]: Total: 10 Nov 27 03:00:45 localhost puppet-user[52669]: Time: Nov 27 03:00:45 localhost puppet-user[52669]: Schedule: 0.00 Nov 27 03:00:45 localhost puppet-user[52669]: File: 0.00 Nov 27 03:00:45 localhost puppet-user[52669]: Augeas: 0.02 Nov 27 03:00:45 localhost puppet-user[52669]: Exec: 0.02 Nov 27 03:00:45 localhost puppet-user[52669]: Transaction evaluation: 0.06 Nov 27 03:00:45 localhost puppet-user[52669]: Catalog application: 0.06 Nov 27 03:00:45 localhost puppet-user[52669]: Config retrieval: 0.17 Nov 27 03:00:45 localhost puppet-user[52669]: Last run: 1764230445 Nov 27 03:00:45 localhost puppet-user[52669]: Filebucket: 0.00 Nov 27 03:00:45 localhost puppet-user[52669]: Total: 0.06 Nov 27 03:00:45 localhost puppet-user[52669]: Version: Nov 27 03:00:45 localhost puppet-user[52669]: Config: 1764230445 Nov 27 03:00:45 localhost puppet-user[52669]: Puppet: 7.10.0 Nov 27 03:00:45 localhost ansible-async_wrapper.py[52650]: Module complete (52650) Nov 27 03:00:45 localhost ansible-async_wrapper.py[52649]: Done in kid B. Nov 27 03:00:51 localhost python3[52875]: ansible-ansible.legacy.async_status Invoked with jid=603233236296.52646 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:00:52 localhost python3[52891]: ansible-file Invoked with path=/var/lib/container-puppet/puppetlabs state=directory setype=svirt_sandbox_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:00:52 localhost python3[52907]: ansible-stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:00:52 localhost python3[52955]: ansible-ansible.legacy.stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:00:53 localhost python3[52998]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/container-puppet/puppetlabs/facter.conf setype=svirt_sandbox_file_t selevel=s0 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230452.5403113-83530-226782667323032/source _original_basename=tmpr6674m6z follow=False checksum=53908622cb869db5e2e2a68e737aa2ab1a872111 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:00:53 localhost python3[53028]: ansible-file Invoked with path=/opt/puppetlabs/facter state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:00:54 localhost python3[53131]: ansible-ansible.posix.synchronize Invoked with src=/opt/puppetlabs/ dest=/var/lib/container-puppet/puppetlabs/ _local_rsync_path=rsync _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] ssh_connection_multiplexing=False partial=False verify_host=False mode=push dest_port=None private_key=None recursive=None links=None perms=None times=None owner=None group=None ssh_args=None link_dest=None Nov 27 03:00:55 localhost python3[53150]: ansible-file Invoked with path=/var/lib/tripleo-config/container-puppet-config mode=448 recurse=True setype=container_file_t force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False state=None _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 03:00:55 localhost python3[53166]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=False puppet_config=/var/lib/container-puppet/container-puppet.json short_hostname=np0005537446 step=1 update_config_hash_only=False Nov 27 03:00:56 localhost python3[53182]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:00:56 localhost python3[53198]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_1 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:00:57 localhost python3[53214]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Nov 27 03:00:57 localhost python3[53256]: ansible-tripleo_container_manage Invoked with config_id=tripleo_puppet_step1 config_dir=/var/lib/tripleo-config/container-puppet-config/step_1 config_patterns=container-puppet-*.json config_overrides={} concurrency=6 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:00:58 localhost podman[53413]: 2025-11-27 08:00:58.285102932 +0000 UTC m=+0.053424828 container create 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_puppet_step1, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=container-puppet-collectd, release=1761123044, batch=17.1_20251118.1, architecture=x86_64, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, managed_by=tripleo_ansible, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:00:58 localhost systemd[1]: Started libpod-conmon-996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59.scope. Nov 27 03:00:58 localhost podman[53447]: 2025-11-27 08:00:58.340275002 +0000 UTC m=+0.070661422 container create 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, io.openshift.expose-services=, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, build-date=2025-11-19T00:35:22Z, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, architecture=x86_64, container_name=container-puppet-nova_libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, tcib_managed=true, name=rhosp17/openstack-nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, release=1761123044, config_id=tripleo_puppet_step1, io.buildah.version=1.41.4, vcs-type=git) Nov 27 03:00:58 localhost systemd[1]: Started libcrun container. Nov 27 03:00:58 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2b0571b203ccc1c35d75f17c46a7652e7cef2ad9c897294271983c8c397fec1d/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:58 localhost podman[53413]: 2025-11-27 08:00:58.354141763 +0000 UTC m=+0.122463649 container init 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, name=rhosp17/openstack-collectd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_puppet_step1, distribution-scope=public, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.openshift.expose-services=, vendor=Red Hat, Inc., container_name=container-puppet-collectd) Nov 27 03:00:58 localhost podman[53413]: 2025-11-27 08:00:58.256278032 +0000 UTC m=+0.024599948 image pull registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 Nov 27 03:00:58 localhost podman[53460]: 2025-11-27 08:00:58.360884999 +0000 UTC m=+0.072414754 container create 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_puppet_step1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=container-puppet-metrics_qdr, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible) Nov 27 03:00:58 localhost systemd[1]: Started libpod-conmon-20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b.scope. Nov 27 03:00:58 localhost systemd[1]: Started libcrun container. Nov 27 03:00:58 localhost systemd[1]: Started libpod-conmon-4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026.scope. Nov 27 03:00:58 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/55e8351b1958513e36671035d0bb47863b4e87c80590aade3f4e58207a1d6315/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:58 localhost systemd[1]: Started libcrun container. Nov 27 03:00:58 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c3e152bf86407d82250d5897387ddaeebc897fdca8787417a054f2cf79ff6916/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:58 localhost podman[53447]: 2025-11-27 08:00:58.311058212 +0000 UTC m=+0.041444642 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:00:58 localhost podman[53460]: 2025-11-27 08:00:58.316756582 +0000 UTC m=+0.028286347 image pull registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 03:00:58 localhost podman[53475]: 2025-11-27 08:00:58.357473515 +0000 UTC m=+0.040742614 image pull registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 Nov 27 03:00:59 localhost podman[53475]: 2025-11-27 08:00:59.013863338 +0000 UTC m=+0.697132457 container create c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, architecture=x86_64, maintainer=OpenStack TripleO Team, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=container-puppet-crond, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_puppet_step1) Nov 27 03:00:59 localhost podman[53413]: 2025-11-27 08:00:59.219169307 +0000 UTC m=+0.987491303 container start 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.expose-services=, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, architecture=x86_64, config_id=tripleo_puppet_step1, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public) Nov 27 03:00:59 localhost podman[53413]: 2025-11-27 08:00:59.221380782 +0000 UTC m=+0.989702768 container attach 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_puppet_step1, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, version=17.1.12, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-type=git, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=container-puppet-collectd, name=rhosp17/openstack-collectd, io.openshift.expose-services=) Nov 27 03:00:59 localhost systemd[1]: Started libpod-conmon-c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2.scope. Nov 27 03:00:59 localhost podman[53447]: 2025-11-27 08:00:59.257603605 +0000 UTC m=+0.987990065 container init 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, config_id=tripleo_puppet_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, version=17.1.12, build-date=2025-11-19T00:35:22Z, name=rhosp17/openstack-nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, container_name=container-puppet-nova_libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044) Nov 27 03:00:59 localhost systemd[1]: Started libcrun container. Nov 27 03:00:59 localhost podman[53460]: 2025-11-27 08:00:59.26312164 +0000 UTC m=+0.974651415 container init 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, release=1761123044, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, config_id=tripleo_puppet_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, url=https://www.redhat.com, vcs-type=git, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=container-puppet-metrics_qdr) Nov 27 03:00:59 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/e5f0de22de697dbbdae069f81567c02a109132469ced70c62116dc3271edb887/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:59 localhost podman[53447]: 2025-11-27 08:00:59.27044538 +0000 UTC m=+1.000831840 container start 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, vcs-type=git, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-libvirt-container, description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=container-puppet-nova_libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_puppet_step1, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-libvirt) Nov 27 03:00:59 localhost podman[53447]: 2025-11-27 08:00:59.271013844 +0000 UTC m=+1.001400334 container attach 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.buildah.version=1.41.4, architecture=x86_64, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, build-date=2025-11-19T00:35:22Z, name=rhosp17/openstack-nova-libvirt, vendor=Red Hat, Inc., config_id=tripleo_puppet_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, batch=17.1_20251118.1, managed_by=tripleo_ansible, container_name=container-puppet-nova_libvirt) Nov 27 03:00:59 localhost podman[53475]: 2025-11-27 08:00:59.275048654 +0000 UTC m=+0.958317773 container init c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, release=1761123044, config_id=tripleo_puppet_step1, container_name=container-puppet-crond, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:00:59 localhost podman[53475]: 2025-11-27 08:00:59.290091564 +0000 UTC m=+0.973360683 container start c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_puppet_step1, summary=Red Hat OpenStack Platform 17.1 cron, container_name=container-puppet-crond, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, vendor=Red Hat, Inc., config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, architecture=x86_64, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:32Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, release=1761123044, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron) Nov 27 03:00:59 localhost podman[53475]: 2025-11-27 08:00:59.29033156 +0000 UTC m=+0.973600679 container attach c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_puppet_step1, vcs-type=git, name=rhosp17/openstack-cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, container_name=container-puppet-crond, distribution-scope=public, tcib_managed=true, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:00:59 localhost podman[53446]: 2025-11-27 08:00:59.319223233 +0000 UTC m=+1.055819246 container create 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, io.buildah.version=1.41.4, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, container_name=container-puppet-iscsid, url=https://www.redhat.com, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_puppet_step1, vcs-type=git, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, distribution-scope=public, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, version=17.1.12, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, architecture=x86_64, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:00:59 localhost podman[53460]: 2025-11-27 08:00:59.323137009 +0000 UTC m=+1.034666784 container start 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, release=1761123044, io.openshift.expose-services=, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-metrics_qdr, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_puppet_step1, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64) Nov 27 03:00:59 localhost podman[53460]: 2025-11-27 08:00:59.323717303 +0000 UTC m=+1.035247128 container attach 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, container_name=container-puppet-metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, architecture=x86_64, io.buildah.version=1.41.4, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_id=tripleo_puppet_step1, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible) Nov 27 03:00:59 localhost podman[53446]: 2025-11-27 08:00:59.246293885 +0000 UTC m=+0.982889948 image pull registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 Nov 27 03:00:59 localhost systemd[1]: Started libpod-conmon-25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9.scope. Nov 27 03:00:59 localhost systemd[1]: Started libcrun container. Nov 27 03:00:59 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/455f92abf31d001bbeee5fb451e818a76066a31777170d227f7983742bedaf10/merged/tmp/iscsi.host supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:59 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/455f92abf31d001bbeee5fb451e818a76066a31777170d227f7983742bedaf10/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:00:59 localhost podman[53446]: 2025-11-27 08:00:59.431041377 +0000 UTC m=+1.167637360 container init 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, version=17.1.12, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, container_name=container-puppet-iscsid, batch=17.1_20251118.1, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, config_id=tripleo_puppet_step1, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:00:59 localhost podman[53446]: 2025-11-27 08:00:59.442487 +0000 UTC m=+1.179083023 container start 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, config_id=tripleo_puppet_step1, container_name=container-puppet-iscsid, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:00:59 localhost podman[53446]: 2025-11-27 08:00:59.444575491 +0000 UTC m=+1.181171514 container attach 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, version=17.1.12, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, url=https://www.redhat.com, container_name=container-puppet-iscsid, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_puppet_step1, architecture=x86_64, managed_by=tripleo_ansible, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, distribution-scope=public, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:01:00 localhost ovs-vsctl[53629]: ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: database connection failed (No such file or directory) Nov 27 03:01:01 localhost puppet-user[53570]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:01 localhost puppet-user[53570]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:01 localhost puppet-user[53570]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:01 localhost puppet-user[53570]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53522]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:01 localhost puppet-user[53522]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:01 localhost puppet-user[53522]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:01 localhost puppet-user[53522]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:01 localhost puppet-user[53564]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:01 localhost puppet-user[53564]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53570]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:01 localhost puppet-user[53570]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:01 localhost puppet-user[53564]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53522]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:01 localhost puppet-user[53522]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53570]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.07 seconds Nov 27 03:01:01 localhost puppet-user[53566]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:01 localhost puppet-user[53566]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:01 localhost puppet-user[53566]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:01 localhost puppet-user[53566]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53566]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:01 localhost puppet-user[53566]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53570]: Notice: /Stage[main]/Tripleo::Profile::Base::Logging::Logrotate/File[/etc/logrotate-crond.conf]/ensure: defined content as '{sha256}1c3202f58bd2ae16cb31badcbb7f0d4e6697157b987d1887736ad96bb73d70b0' Nov 27 03:01:01 localhost puppet-user[53570]: Notice: /Stage[main]/Tripleo::Profile::Base::Logging::Logrotate/Cron[logrotate-crond]/ensure: created Nov 27 03:01:01 localhost puppet-user[53570]: Notice: Applied catalog in 0.04 seconds Nov 27 03:01:01 localhost puppet-user[53570]: Application: Nov 27 03:01:01 localhost puppet-user[53570]: Initial environment: production Nov 27 03:01:01 localhost puppet-user[53570]: Converged environment: production Nov 27 03:01:01 localhost puppet-user[53570]: Run mode: user Nov 27 03:01:01 localhost puppet-user[53570]: Changes: Nov 27 03:01:01 localhost puppet-user[53570]: Total: 2 Nov 27 03:01:01 localhost puppet-user[53570]: Events: Nov 27 03:01:01 localhost puppet-user[53570]: Success: 2 Nov 27 03:01:01 localhost puppet-user[53570]: Total: 2 Nov 27 03:01:01 localhost puppet-user[53570]: Resources: Nov 27 03:01:01 localhost puppet-user[53570]: Changed: 2 Nov 27 03:01:01 localhost puppet-user[53570]: Out of sync: 2 Nov 27 03:01:01 localhost puppet-user[53570]: Skipped: 7 Nov 27 03:01:01 localhost puppet-user[53570]: Total: 9 Nov 27 03:01:01 localhost puppet-user[53570]: Time: Nov 27 03:01:01 localhost puppet-user[53570]: File: 0.00 Nov 27 03:01:01 localhost puppet-user[53570]: Cron: 0.01 Nov 27 03:01:01 localhost puppet-user[53570]: Transaction evaluation: 0.03 Nov 27 03:01:01 localhost puppet-user[53570]: Catalog application: 0.04 Nov 27 03:01:01 localhost puppet-user[53570]: Config retrieval: 0.11 Nov 27 03:01:01 localhost puppet-user[53570]: Last run: 1764230461 Nov 27 03:01:01 localhost puppet-user[53570]: Total: 0.04 Nov 27 03:01:01 localhost puppet-user[53570]: Version: Nov 27 03:01:01 localhost puppet-user[53570]: Config: 1764230461 Nov 27 03:01:01 localhost puppet-user[53570]: Puppet: 7.10.0 Nov 27 03:01:01 localhost puppet-user[53566]: Notice: Accepting previously invalid value for target type 'Integer' Nov 27 03:01:01 localhost puppet-user[53566]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.12 seconds Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/var/lib/qdrouterd]/owner: owner changed 'qdrouterd' to 'root' Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/var/lib/qdrouterd]/group: group changed 'qdrouterd' to 'root' Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/var/lib/qdrouterd]/mode: mode changed '0700' to '0755' Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/etc/qpid-dispatch/ssl]/ensure: created Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[qdrouterd.conf]/content: content changed '{sha256}89e10d8896247f992c5f0baf027c25a8ca5d0441be46d8859d9db2067ea74cd3' to '{sha256}c4eb9a3c8bbde7ca21bccc30e19d82e283e4fc12558f6f9a14980caff2c996ff' Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/var/log/qdrouterd]/ensure: created Nov 27 03:01:01 localhost puppet-user[53566]: Notice: /Stage[main]/Qdr::Config/File[/var/log/qdrouterd/metrics_qdr.log]/ensure: created Nov 27 03:01:01 localhost puppet-user[53566]: Notice: Applied catalog in 0.03 seconds Nov 27 03:01:01 localhost puppet-user[53566]: Application: Nov 27 03:01:01 localhost puppet-user[53566]: Initial environment: production Nov 27 03:01:01 localhost puppet-user[53566]: Converged environment: production Nov 27 03:01:01 localhost puppet-user[53566]: Run mode: user Nov 27 03:01:01 localhost puppet-user[53566]: Changes: Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Nova]): The os_region_name parameter is deprecated and will be removed \ Nov 27 03:01:01 localhost puppet-user[53564]: in a future release. Use nova::cinder::os_region_name instead Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Nova]): The catalog_info parameter is deprecated and will be removed \ Nov 27 03:01:01 localhost puppet-user[53564]: in a future release. Use nova::cinder::catalog_info instead Nov 27 03:01:01 localhost puppet-user[53566]: Total: 7 Nov 27 03:01:01 localhost puppet-user[53566]: Events: Nov 27 03:01:01 localhost puppet-user[53566]: Success: 7 Nov 27 03:01:01 localhost puppet-user[53566]: Total: 7 Nov 27 03:01:01 localhost puppet-user[53566]: Resources: Nov 27 03:01:01 localhost puppet-user[53566]: Skipped: 13 Nov 27 03:01:01 localhost puppet-user[53566]: Changed: 5 Nov 27 03:01:01 localhost puppet-user[53566]: Out of sync: 5 Nov 27 03:01:01 localhost puppet-user[53566]: Total: 20 Nov 27 03:01:01 localhost puppet-user[53566]: Time: Nov 27 03:01:01 localhost puppet-user[53566]: File: 0.01 Nov 27 03:01:01 localhost puppet-user[53566]: Transaction evaluation: 0.03 Nov 27 03:01:01 localhost puppet-user[53566]: Catalog application: 0.03 Nov 27 03:01:01 localhost puppet-user[53566]: Config retrieval: 0.15 Nov 27 03:01:01 localhost puppet-user[53566]: Last run: 1764230461 Nov 27 03:01:01 localhost puppet-user[53566]: Total: 0.03 Nov 27 03:01:01 localhost puppet-user[53566]: Version: Nov 27 03:01:01 localhost puppet-user[53566]: Config: 1764230461 Nov 27 03:01:01 localhost puppet-user[53566]: Puppet: 7.10.0 Nov 27 03:01:01 localhost puppet-user[53522]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.35 seconds Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Unknown variable: '::nova::compute::verify_glance_signatures'. (file: /etc/puppet/modules/nova/manifests/glance.pp, line: 62, column: 41) Nov 27 03:01:01 localhost puppet-user[53591]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:01 localhost puppet-user[53591]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:01 localhost puppet-user[53591]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:01 localhost puppet-user[53591]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53591]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:01 localhost puppet-user[53591]: (file & line not available) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Unknown variable: '::nova::compute::libvirt::remove_unused_base_images'. (file: /etc/puppet/modules/nova/manifests/compute/image_cache.pp, line: 44, column: 5) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Unknown variable: '::nova::compute::libvirt::remove_unused_original_minimum_age_seconds'. (file: /etc/puppet/modules/nova/manifests/compute/image_cache.pp, line: 48, column: 5) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Unknown variable: '::nova::compute::libvirt::remove_unused_resized_minimum_age_seconds'. (file: /etc/puppet/modules/nova/manifests/compute/image_cache.pp, line: 52, column: 5) Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Tripleo::Profile::Base::Nova::Compute]): The keymgr_backend parameter has been deprecated Nov 27 03:01:01 localhost systemd[1]: libpod-c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2.scope: Deactivated successfully. Nov 27 03:01:01 localhost systemd[1]: libpod-c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2.scope: Consumed 2.129s CPU time. Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Nova::Compute]): vcpu_pin_set is deprecated, instead use cpu_dedicated_set or cpu_shared_set. Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Nova::Compute]): verify_glance_signatures is deprecated. Use the same parameter in nova::glance Nov 27 03:01:01 localhost puppet-user[53591]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.14 seconds Nov 27 03:01:01 localhost podman[53475]: 2025-11-27 08:01:01.562111685 +0000 UTC m=+3.245380784 container died c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-crond, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, com.redhat.component=openstack-cron-container, vcs-type=git, config_id=tripleo_puppet_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:01:01 localhost puppet-user[53591]: Notice: /Stage[main]/Tripleo::Profile::Base::Iscsid/Exec[reset-iscsi-initiator-name]/returns: executed successfully Nov 27 03:01:01 localhost puppet-user[53591]: Notice: /Stage[main]/Tripleo::Profile::Base::Iscsid/File[/etc/iscsi/.initiator_reset]/ensure: created Nov 27 03:01:01 localhost systemd[1]: tmp-crun.AryFZe.mount: Deactivated successfully. Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.conf]/content: content changed '{sha256}aea388a73ebafc7e07a81ddb930a91099211f660eee55fbf92c13007a77501e5' to '{sha256}2523d01ee9c3022c0e9f61d896b1474a168e18472aee141cc278e69fe13f41c1' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.conf]/owner: owner changed 'collectd' to 'root' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.conf]/group: group changed 'collectd' to 'root' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.conf]/mode: mode changed '0644' to '0640' Nov 27 03:01:01 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:01 localhost systemd[1]: var-lib-containers-storage-overlay-e5f0de22de697dbbdae069f81567c02a109132469ced70c62116dc3271edb887-merged.mount: Deactivated successfully. Nov 27 03:01:01 localhost puppet-user[53591]: Notice: /Stage[main]/Tripleo::Profile::Base::Iscsid/Exec[sync-iqn-to-host]/returns: executed successfully Nov 27 03:01:01 localhost systemd[1]: libpod-4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026.scope: Deactivated successfully. Nov 27 03:01:01 localhost systemd[1]: libpod-4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026.scope: Consumed 2.223s CPU time. Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.d]/owner: owner changed 'collectd' to 'root' Nov 27 03:01:01 localhost podman[53460]: 2025-11-27 08:01:01.685869394 +0000 UTC m=+3.397399139 container died 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=container-puppet-metrics_qdr, config_id=tripleo_puppet_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, release=1761123044, vendor=Red Hat, Inc., tcib_managed=true, io.openshift.expose-services=, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, architecture=x86_64, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.d]/group: group changed 'collectd' to 'root' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[collectd.d]/mode: mode changed '0755' to '0750' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/90-default-plugins-cpu.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/90-default-plugins-interface.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/90-default-plugins-load.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/90-default-plugins-memory.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/90-default-plugins-syslog.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/apache.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/dns.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/ipmi.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/mcelog.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/mysql.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/ovs-events.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/ovs-stats.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/ping.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/pmu.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/rdt.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/sensors.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/snmp.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Config/File[/etc/collectd.d/write_prometheus.conf]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Python/File[/usr/lib/python3.9/site-packages]/mode: mode changed '0755' to '0750' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Python/Collectd::Plugin[python]/File[python.load]/ensure: defined content as '{sha256}0163924a0099dd43fe39cb85e836df147fd2cfee8197dc6866d3c384539eb6ee' Nov 27 03:01:01 localhost podman[53989]: 2025-11-27 08:01:01.703979681 +0000 UTC m=+0.133038089 container cleanup c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2 (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=container-puppet-crond, io.openshift.expose-services=, distribution-scope=public, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=container-puppet-crond, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, config_id=tripleo_puppet_step1, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Python/Concat[/etc/collectd.d/python-config.conf]/File[/etc/collectd.d/python-config.conf]/ensure: defined content as '{sha256}2e5fb20e60b30f84687fc456a37fc62451000d2d85f5bbc1b3fca3a5eac9deeb' Nov 27 03:01:01 localhost systemd[1]: libpod-conmon-c66738732109cebf4ba90c497a1783930831b99a2dc4631be5fe064034718bf2.scope: Deactivated successfully. Nov 27 03:01:01 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-crond --conmon-pidfile /run/container-puppet-crond.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron --env NAME=crond --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::logging::logrotate --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-crond --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'crond', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::logrotate'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-crond.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Logfile/Collectd::Plugin[logfile]/File[logfile.load]/ensure: defined content as '{sha256}07bbda08ef9b824089500bdc6ac5a86e7d1ef2ae3ed4ed423c0559fe6361e5af' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Amqp1/Collectd::Plugin[amqp1]/File[amqp1.load]/ensure: defined content as '{sha256}8dd3769945b86c38433504b97f7851a931eb3c94b667298d10a9796a3d020595' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Ceph/Collectd::Plugin[ceph]/File[ceph.load]/ensure: defined content as '{sha256}c796abffda2e860875295b4fc11cc95c6032b4e13fa8fb128e839a305aa1676c' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Cpu/Collectd::Plugin[cpu]/File[cpu.load]/ensure: defined content as '{sha256}67d4c8bf6bf5785f4cb6b596712204d9eacbcebbf16fe289907195d4d3cb0e34' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Df/Collectd::Plugin[df]/File[df.load]/ensure: defined content as '{sha256}edeb4716d96fc9dca2c6adfe07bae70ba08c6af3944a3900581cba0f08f3c4ba' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Disk/Collectd::Plugin[disk]/File[disk.load]/ensure: defined content as '{sha256}1d0cb838278f3226fcd381f0fc2e0e1abaf0d590f4ba7bcb2fc6ec113d3ebde7' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Hugepages/Collectd::Plugin[hugepages]/File[hugepages.load]/ensure: defined content as '{sha256}9b9f35b65a73da8d4037e4355a23b678f2cf61997ccf7a5e1adf2a7ce6415827' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Hugepages/Collectd::Plugin[hugepages]/File[older_hugepages.load]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Interface/Collectd::Plugin[interface]/File[interface.load]/ensure: defined content as '{sha256}b76b315dc312e398940fe029c6dbc5c18d2b974ff7527469fc7d3617b5222046' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Load/Collectd::Plugin[load]/File[load.load]/ensure: defined content as '{sha256}af2403f76aebd2f10202d66d2d55e1a8d987eed09ced5a3e3873a4093585dc31' Nov 27 03:01:01 localhost puppet-user[53564]: Warning: Scope(Class[Nova::Compute::Libvirt]): nova::compute::libvirt::images_type will be required if rbd ephemeral storage is used. Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Memory/Collectd::Plugin[memory]/File[memory.load]/ensure: defined content as '{sha256}0f270425ee6b05fc9440ee32b9afd1010dcbddd9b04ca78ff693858f7ecb9d0e' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Unixsock/Collectd::Plugin[unixsock]/File[unixsock.load]/ensure: defined content as '{sha256}9d1ec1c51ba386baa6f62d2e019dbd6998ad924bf868b3edc2d24d3dc3c63885' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Uptime/Collectd::Plugin[uptime]/File[uptime.load]/ensure: defined content as '{sha256}f7a26c6369f904d0ca1af59627ebea15f5e72160bcacdf08d217af282b42e5c0' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Virt/Collectd::Plugin[virt]/File[virt.load]/ensure: defined content as '{sha256}9a2bcf913f6bf8a962a0ff351a9faea51ae863cc80af97b77f63f8ab68941c62' Nov 27 03:01:01 localhost puppet-user[53522]: Notice: /Stage[main]/Collectd::Plugin::Virt/Collectd::Plugin[virt]/File[older_virt.load]/ensure: removed Nov 27 03:01:01 localhost puppet-user[53522]: Notice: Applied catalog in 0.19 seconds Nov 27 03:01:01 localhost puppet-user[53522]: Application: Nov 27 03:01:01 localhost puppet-user[53522]: Initial environment: production Nov 27 03:01:01 localhost puppet-user[53522]: Converged environment: production Nov 27 03:01:01 localhost puppet-user[53522]: Run mode: user Nov 27 03:01:01 localhost puppet-user[53522]: Changes: Nov 27 03:01:01 localhost puppet-user[53522]: Total: 43 Nov 27 03:01:01 localhost puppet-user[53522]: Events: Nov 27 03:01:01 localhost puppet-user[53522]: Success: 43 Nov 27 03:01:01 localhost puppet-user[53522]: Total: 43 Nov 27 03:01:01 localhost puppet-user[53522]: Resources: Nov 27 03:01:01 localhost puppet-user[53522]: Skipped: 14 Nov 27 03:01:01 localhost puppet-user[53522]: Changed: 38 Nov 27 03:01:01 localhost puppet-user[53522]: Out of sync: 38 Nov 27 03:01:01 localhost puppet-user[53522]: Total: 82 Nov 27 03:01:01 localhost puppet-user[53522]: Time: Nov 27 03:01:01 localhost puppet-user[53522]: Concat file: 0.00 Nov 27 03:01:01 localhost puppet-user[53522]: File: 0.08 Nov 27 03:01:01 localhost puppet-user[53522]: Transaction evaluation: 0.18 Nov 27 03:01:01 localhost puppet-user[53522]: Catalog application: 0.19 Nov 27 03:01:01 localhost puppet-user[53522]: Config retrieval: 0.53 Nov 27 03:01:01 localhost puppet-user[53522]: Last run: 1764230461 Nov 27 03:01:01 localhost puppet-user[53522]: Concat fragment: 0.00 Nov 27 03:01:01 localhost puppet-user[53522]: Total: 0.19 Nov 27 03:01:01 localhost puppet-user[53522]: Version: Nov 27 03:01:01 localhost puppet-user[53522]: Config: 1764230461 Nov 27 03:01:01 localhost puppet-user[53522]: Puppet: 7.10.0 Nov 27 03:01:01 localhost podman[54030]: 2025-11-27 08:01:01.904110492 +0000 UTC m=+0.205787281 container cleanup 4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=container-puppet-metrics_qdr, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container, release=1761123044, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_puppet_step1, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=container-puppet-metrics_qdr, tcib_managed=true, distribution-scope=public, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, url=https://www.redhat.com, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 03:01:01 localhost systemd[1]: libpod-conmon-4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026.scope: Deactivated successfully. Nov 27 03:01:01 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-metrics_qdr --conmon-pidfile /run/container-puppet-metrics_qdr.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron --env NAME=metrics_qdr --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::metrics::qdr#012 --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-metrics_qdr --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron', 'NAME': 'metrics_qdr', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::qdr\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-metrics_qdr.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 03:01:02 localhost puppet-user[53591]: Notice: /Stage[main]/Tripleo::Profile::Base::Iscsid/Augeas[chap_algs in /etc/iscsi/iscsid.conf]/returns: executed successfully Nov 27 03:01:02 localhost puppet-user[53591]: Notice: Applied catalog in 0.47 seconds Nov 27 03:01:02 localhost puppet-user[53591]: Application: Nov 27 03:01:02 localhost puppet-user[53591]: Initial environment: production Nov 27 03:01:02 localhost puppet-user[53591]: Converged environment: production Nov 27 03:01:02 localhost puppet-user[53591]: Run mode: user Nov 27 03:01:02 localhost puppet-user[53591]: Changes: Nov 27 03:01:02 localhost puppet-user[53591]: Total: 4 Nov 27 03:01:02 localhost puppet-user[53591]: Events: Nov 27 03:01:02 localhost puppet-user[53591]: Success: 4 Nov 27 03:01:02 localhost puppet-user[53591]: Total: 4 Nov 27 03:01:02 localhost puppet-user[53591]: Resources: Nov 27 03:01:02 localhost puppet-user[53591]: Changed: 4 Nov 27 03:01:02 localhost puppet-user[53591]: Out of sync: 4 Nov 27 03:01:02 localhost puppet-user[53591]: Skipped: 8 Nov 27 03:01:02 localhost puppet-user[53591]: Total: 13 Nov 27 03:01:02 localhost puppet-user[53591]: Time: Nov 27 03:01:02 localhost puppet-user[53591]: File: 0.00 Nov 27 03:01:02 localhost puppet-user[53591]: Exec: 0.06 Nov 27 03:01:02 localhost puppet-user[53591]: Config retrieval: 0.18 Nov 27 03:01:02 localhost puppet-user[53591]: Augeas: 0.39 Nov 27 03:01:02 localhost puppet-user[53591]: Transaction evaluation: 0.46 Nov 27 03:01:02 localhost puppet-user[53591]: Catalog application: 0.47 Nov 27 03:01:02 localhost puppet-user[53591]: Last run: 1764230462 Nov 27 03:01:02 localhost puppet-user[53591]: Total: 0.47 Nov 27 03:01:02 localhost puppet-user[53591]: Version: Nov 27 03:01:02 localhost puppet-user[53591]: Config: 1764230461 Nov 27 03:01:02 localhost puppet-user[53591]: Puppet: 7.10.0 Nov 27 03:01:02 localhost systemd[1]: libpod-996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59.scope: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: libpod-996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59.scope: Consumed 2.776s CPU time. Nov 27 03:01:02 localhost podman[53413]: 2025-11-27 08:01:02.160191752 +0000 UTC m=+3.928513658 container died 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, name=rhosp17/openstack-collectd, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_puppet_step1, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.openshift.expose-services=, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, vcs-type=git, container_name=container-puppet-collectd, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true) Nov 27 03:01:02 localhost podman[54152]: 2025-11-27 08:01:02.172943166 +0000 UTC m=+0.083078008 container create fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, vendor=Red Hat, Inc., com.redhat.component=openstack-rsyslog-container, io.openshift.expose-services=, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_puppet_step1, version=17.1.12, vcs-type=git, distribution-scope=public, tcib_managed=true, name=rhosp17/openstack-rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.buildah.version=1.41.4, container_name=container-puppet-rsyslog, release=1761123044, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 rsyslog, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, build-date=2025-11-18T22:49:49Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:01:02 localhost systemd[1]: Started libpod-conmon-fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586.scope. Nov 27 03:01:02 localhost podman[54152]: 2025-11-27 08:01:02.125300112 +0000 UTC m=+0.035435024 image pull registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 Nov 27 03:01:02 localhost systemd[1]: Started libcrun container. Nov 27 03:01:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0034e3c51dba6c91c611dd6ead3f9e70360969b5749ffdec598d74f5bdf45edb/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:02 localhost podman[54152]: 2025-11-27 08:01:02.301604366 +0000 UTC m=+0.211739208 container init fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-rsyslog, distribution-scope=public, name=rhosp17/openstack-rsyslog, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.component=openstack-rsyslog-container, managed_by=tripleo_ansible, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.expose-services=, config_id=tripleo_puppet_step1, release=1761123044, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, summary=Red Hat OpenStack Platform 17.1 rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:49Z, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog) Nov 27 03:01:02 localhost podman[54152]: 2025-11-27 08:01:02.314157346 +0000 UTC m=+0.224292188 container start fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, batch=17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, build-date=2025-11-18T22:49:49Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.component=openstack-rsyslog-container, architecture=x86_64, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, name=rhosp17/openstack-rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, distribution-scope=public, version=17.1.12, container_name=container-puppet-rsyslog, config_id=tripleo_puppet_step1, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog) Nov 27 03:01:02 localhost podman[54152]: 2025-11-27 08:01:02.314407672 +0000 UTC m=+0.224542534 container attach fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, name=rhosp17/openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, maintainer=OpenStack TripleO Team, release=1761123044, architecture=x86_64, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, distribution-scope=public, config_id=tripleo_puppet_step1, url=https://www.redhat.com, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.component=openstack-rsyslog-container, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.openshift.expose-services=, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, container_name=container-puppet-rsyslog) Nov 27 03:01:02 localhost podman[54191]: 2025-11-27 08:01:02.319017405 +0000 UTC m=+0.148145922 container cleanup 996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=container-puppet-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, container_name=container-puppet-collectd, config_id=tripleo_puppet_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-type=git, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., tcib_managed=true) Nov 27 03:01:02 localhost systemd[1]: libpod-conmon-996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59.scope: Deactivated successfully. Nov 27 03:01:02 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-collectd --conmon-pidfile /run/container-puppet-collectd.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,collectd_client_config,exec --env NAME=collectd --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::metrics::collectd --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-collectd --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,collectd_client_config,exec', 'NAME': 'collectd', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::metrics::collectd'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-collectd.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 Nov 27 03:01:02 localhost puppet-user[53564]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 1.29 seconds Nov 27 03:01:02 localhost systemd[1]: libpod-25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9.scope: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: libpod-25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9.scope: Consumed 2.804s CPU time. Nov 27 03:01:02 localhost podman[54232]: 2025-11-27 08:01:02.373948369 +0000 UTC m=+0.086326459 container create 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, distribution-scope=public, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, tcib_managed=true, io.buildah.version=1.41.4, architecture=x86_64, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, config_id=tripleo_puppet_step1, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=container-puppet-ovn_controller, release=1761123044, vcs-type=git) Nov 27 03:01:02 localhost systemd[1]: Started libpod-conmon-2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b.scope. Nov 27 03:01:02 localhost systemd[1]: Started libcrun container. Nov 27 03:01:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/973a1ad437b898f3b247f823900f22fbefb6e9a3e82863653f54b4dbb77d854c/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/973a1ad437b898f3b247f823900f22fbefb6e9a3e82863653f54b4dbb77d854c/merged/etc/sysconfig/modules supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:02 localhost podman[54232]: 2025-11-27 08:01:02.319566739 +0000 UTC m=+0.031944839 image pull registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:01:02 localhost podman[54232]: 2025-11-27 08:01:02.42105914 +0000 UTC m=+0.133437260 container init 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, container_name=container-puppet-ovn_controller, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, architecture=x86_64, distribution-scope=public, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, tcib_managed=true, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, config_id=tripleo_puppet_step1, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z) Nov 27 03:01:02 localhost podman[53446]: 2025-11-27 08:01:02.425836737 +0000 UTC m=+4.162432750 container died 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.expose-services=, container_name=container-puppet-iscsid, build-date=2025-11-18T23:44:13Z, architecture=x86_64, release=1761123044, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_puppet_step1, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, vcs-type=git, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:01:02 localhost podman[54232]: 2025-11-27 08:01:02.482191426 +0000 UTC m=+0.194569546 container start 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=container-puppet-ovn_controller, name=rhosp17/openstack-ovn-controller, io.buildah.version=1.41.4, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, tcib_managed=true, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, config_id=tripleo_puppet_step1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:01:02 localhost podman[54232]: 2025-11-27 08:01:02.482455942 +0000 UTC m=+0.194834062 container attach 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, config_id=tripleo_puppet_step1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=container-puppet-ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, vendor=Red Hat, Inc., managed_by=tripleo_ansible) Nov 27 03:01:02 localhost podman[54292]: 2025-11-27 08:01:02.499293167 +0000 UTC m=+0.111705873 container cleanup 25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=container-puppet-iscsid, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, architecture=x86_64, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., container_name=container-puppet-iscsid, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, config_id=tripleo_puppet_step1, tcib_managed=true, vcs-type=git) Nov 27 03:01:02 localhost systemd[1]: libpod-conmon-25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9.scope: Deactivated successfully. Nov 27 03:01:02 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-iscsid --conmon-pidfile /run/container-puppet-iscsid.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,iscsid_config --env NAME=iscsid --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::iscsid#012 --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-iscsid --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,iscsid_config', 'NAME': 'iscsid', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::iscsid\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/iscsi:/tmp/iscsi.host:z', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-iscsid.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/iscsi:/tmp/iscsi.host:z --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay-455f92abf31d001bbeee5fb451e818a76066a31777170d227f7983742bedaf10-merged.mount: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-25d41a5df15eb252f957ff008a35dc72085e90e4b4b64dd44dd8515105d5b3b9-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay-c3e152bf86407d82250d5897387ddaeebc897fdca8787417a054f2cf79ff6916-merged.mount: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-4e71e885680b3f71ef65119aa83fdf586d34ca8ac5d7141801fee95c12820026-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay-2b0571b203ccc1c35d75f17c46a7652e7cef2ad9c897294271983c8c397fec1d-merged.mount: Deactivated successfully. Nov 27 03:01:02 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-996f220ea8a015c93d1572c87ceab8f76da26f6115c1854fb212f9fe6e64cd59-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Migration::Client/File[/etc/nova/migration/identity]/content: content changed '{sha256}86610d84e745a3992358ae0b747297805d075492e5114c666fa08f8aecce7da0' to '{sha256}23bf6829192f896c3f1a4a2238962441835ac3bebe043fc5b0777117267f4063' Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Migration::Client/File_line[nova_ssh_port]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Libvirt/File[/etc/sasl2/libvirt.conf]/content: content changed '{sha256}78510a0d6f14b269ddeb9f9638dfdfba9f976d370ee2ec04ba25352a8af6df35' to '{sha256}6d7bcae773217a30c0772f75d0d1b6d21f5d64e72853f5e3d91bb47799dbb7fe' Nov 27 03:01:02 localhost puppet-user[53564]: Warning: Empty environment setting 'TLS_PASSWORD' Nov 27 03:01:02 localhost puppet-user[53564]: (file: /etc/puppet/modules/tripleo/manifests/profile/base/nova/libvirt.pp, line: 182) Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Libvirt/Exec[set libvirt sasl credentials]/returns: executed successfully Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Migration::Target/File[/etc/nova/migration/authorized_keys]/content: content changed '{sha256}0d05a8832f36c0517b84e9c3ad11069d531c7d2be5297661e5552fd29e3a5e47' to '{sha256}7738843a4029bba3465722cb5f7f235d96982cd692b768632f7dbb2a32d70964' Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Tripleo::Profile::Base::Nova::Migration::Target/File_line[nova_migration_logindefs]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Workarounds/Nova_config[workarounds/never_download_image_if_on_rbd]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Workarounds/Nova_config[workarounds/disable_compute_service_check_for_ffu]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/ssl_only]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/my_ip]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/host]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/cpu_allocation_ratio]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/ram_allocation_ratio]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/disk_allocation_ratio]/ensure: created Nov 27 03:01:02 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/dhcp_domain]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[vif_plug_ovs/ovsdb_connection]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[notifications/notification_format]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/state_path]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/service_down_time]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/rootwrap_config]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[DEFAULT/report_interval]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[notifications/notify_on_state_change]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Nova_config[cinder/cross_az_attach]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Glance/Nova_config[glance/valid_interfaces]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/auth_type]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/auth_url]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/password]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/project_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/project_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/user_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/username]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/region_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Placement/Nova_config[placement/valid_interfaces]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/password]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/auth_type]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/auth_url]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/region_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/project_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/project_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/username]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/user_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/os_region_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cinder/Nova_config[cinder/catalog_info]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Image_cache/Nova_config[image_cache/manager_interval]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Image_cache/Nova_config[image_cache/remove_unused_base_images]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Image_cache/Nova_config[image_cache/remove_unused_original_minimum_age_seconds]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Image_cache/Nova_config[image_cache/remove_unused_resized_minimum_age_seconds]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Image_cache/Nova_config[image_cache/precache_concurrency]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Vendordata/Nova_config[vendordata_dynamic_auth/project_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Vendordata/Nova_config[vendordata_dynamic_auth/user_domain_name]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Provider/Nova_config[compute/provider_config_location]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Provider/File[/etc/nova/provider_config]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/use_cow_images]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/mkisofs_cmd]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/force_raw_images]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/reserved_host_memory_mb]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/reserved_huge_pages]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/resume_guests_state_on_host_boot]/ensure: created Nov 27 03:01:03 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[key_manager/backend]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/sync_power_state_interval]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[compute/consecutive_build_service_disable_threshold]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[compute/live_migration_wait_for_vif_plug]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[compute/max_disk_devices_to_attach]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Vncproxy::Common/Nova_config[vnc/novncproxy_base_url]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[vnc/server_proxyclient_address]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[vnc/enabled]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[spice/enabled]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/instance_usage_audit]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute/Nova_config[DEFAULT/instance_usage_audit_period]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[DEFAULT/vif_plugging_is_fatal]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[DEFAULT/vif_plugging_timeout]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/default_floating_pool]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/timeout]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/project_name]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/project_domain_name]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/region_name]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/username]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/user_domain_name]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/password]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/auth_url]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/valid_interfaces]/ensure: created Nov 27 03:01:04 localhost puppet-user[54282]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:04 localhost puppet-user[54282]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:04 localhost puppet-user[54282]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:04 localhost puppet-user[54282]: (file & line not available) Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/ovs_bridge]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/extension_sync_interval]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Network::Neutron/Nova_config[neutron/auth_type]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:04 localhost puppet-user[54320]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:04 localhost puppet-user[54320]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:04 localhost puppet-user[54320]: (file & line not available) Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Nova_config[libvirt/live_migration_uri]/ensure: created Nov 27 03:01:04 localhost puppet-user[54282]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:04 localhost puppet-user[54282]: (file & line not available) Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Nova_config[libvirt/live_migration_tunnelled]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:04 localhost puppet-user[54320]: (file & line not available) Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Nova_config[libvirt/live_migration_inbound_addr]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Nova_config[libvirt/live_migration_permit_post_copy]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Nova_config[libvirt/live_migration_permit_auto_converge]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Virtproxyd_config[listen_tls]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Libvirt/Virtproxyd_config[listen_tcp]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/rbd_user]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/rbd_secret_uuid]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/File[/etc/nova/secret.xml]/ensure: defined content as '{sha256}c590812288546ef605466c696f7b052c4a792fbb3c2e47670e5fc6211e7156a4' Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_type]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_rbd_pool]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_rbd_ceph_conf]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_rbd_glance_store_name]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_rbd_glance_copy_poll_interval]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Rbd/Nova_config[libvirt/images_rbd_glance_copy_timeout]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[DEFAULT/compute_driver]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[DEFAULT/preallocate_images]/ensure: created Nov 27 03:01:04 localhost puppet-user[54282]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.26 seconds Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[vnc/server_listen]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/virt_type]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/cpu_mode]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.26 seconds Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/inject_password]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/inject_key]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/inject_partition]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/hw_disk_discard]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/hw_machine_type]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/enabled_perf_events]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/rx_queue_size]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/tx_queue_size]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54518]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-remote=tcp:172.17.0.103:6642,tcp:172.17.0.104:6642,tcp:172.17.0.105:6642 Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-remote]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/file_backed_memory]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/volume_use_multipath]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54520]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-encap-type=geneve Nov 27 03:01:04 localhost puppet-user[54282]: Notice: /Stage[main]/Rsyslog::Base/File[/etc/rsyslog.conf]/content: content changed '{sha256}d6f679f6a4eb6f33f9fc20c846cb30bef93811e1c86bc4da1946dc3100b826c3' to '{sha256}7963bd801fadd49a17561f4d3f80738c3f504b413b11c443432d8303138041f2' Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/num_pcie_ports]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-encap-type]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/mem_stats_period_seconds]/ensure: created Nov 27 03:01:04 localhost puppet-user[54282]: Notice: /Stage[main]/Rsyslog::Config::Global/Rsyslog::Component::Global_config[MaxMessageSize]/Rsyslog::Generate_concat[rsyslog::concat::global_config::MaxMessageSize]/Concat[/etc/rsyslog.d/00_rsyslog.conf]/File[/etc/rsyslog.d/00_rsyslog.conf]/ensure: defined content as '{sha256}a291d5cc6d5884a978161f4c7b5831d43edd07797cc590bae366e7f150b8643b' Nov 27 03:01:04 localhost ovs-vsctl[54522]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-encap-ip=172.19.0.108 Nov 27 03:01:04 localhost puppet-user[54282]: Notice: /Stage[main]/Rsyslog::Config::Templates/Rsyslog::Component::Template[rsyslog-node-index]/Rsyslog::Generate_concat[rsyslog::concat::template::rsyslog-node-index]/Concat[/etc/rsyslog.d/50_openstack_logs.conf]/File[/etc/rsyslog.d/50_openstack_logs.conf]/ensure: defined content as '{sha256}f443f4220d9a63a6ca6feda5638e4702d403812ebd51f48abfaf70745b67252d' Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/pmem_namespaces]/ensure: created Nov 27 03:01:04 localhost puppet-user[54282]: Notice: Applied catalog in 0.14 seconds Nov 27 03:01:04 localhost puppet-user[54282]: Application: Nov 27 03:01:04 localhost puppet-user[54282]: Initial environment: production Nov 27 03:01:04 localhost puppet-user[54282]: Converged environment: production Nov 27 03:01:04 localhost puppet-user[54282]: Run mode: user Nov 27 03:01:04 localhost puppet-user[54282]: Changes: Nov 27 03:01:04 localhost puppet-user[54282]: Total: 3 Nov 27 03:01:04 localhost puppet-user[54282]: Events: Nov 27 03:01:04 localhost puppet-user[54282]: Success: 3 Nov 27 03:01:04 localhost puppet-user[54282]: Total: 3 Nov 27 03:01:04 localhost puppet-user[54282]: Resources: Nov 27 03:01:04 localhost puppet-user[54282]: Skipped: 11 Nov 27 03:01:04 localhost puppet-user[54282]: Changed: 3 Nov 27 03:01:04 localhost puppet-user[54282]: Out of sync: 3 Nov 27 03:01:04 localhost puppet-user[54282]: Total: 25 Nov 27 03:01:04 localhost puppet-user[54282]: Time: Nov 27 03:01:04 localhost puppet-user[54282]: Concat file: 0.00 Nov 27 03:01:04 localhost puppet-user[54282]: Concat fragment: 0.00 Nov 27 03:01:04 localhost puppet-user[54282]: File: 0.02 Nov 27 03:01:04 localhost puppet-user[54282]: Transaction evaluation: 0.14 Nov 27 03:01:04 localhost puppet-user[54282]: Catalog application: 0.14 Nov 27 03:01:04 localhost puppet-user[54282]: Config retrieval: 0.30 Nov 27 03:01:04 localhost puppet-user[54282]: Last run: 1764230464 Nov 27 03:01:04 localhost puppet-user[54282]: Total: 0.14 Nov 27 03:01:04 localhost puppet-user[54282]: Version: Nov 27 03:01:04 localhost puppet-user[54282]: Config: 1764230464 Nov 27 03:01:04 localhost puppet-user[54282]: Puppet: 7.10.0 Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-encap-ip]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/swtpm_enabled]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54525]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:hostname=np0005537446.localdomain Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:hostname]/value: value changed 'np0005537446.novalocal' to 'np0005537446.localdomain' Nov 27 03:01:04 localhost ovs-vsctl[54527]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-bridge=br-int Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-bridge]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/cpu_model_extra_flags]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt/Nova_config[libvirt/disk_cachemodes]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtlogd/Virtlogd_config[log_filters]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtlogd/Virtlogd_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtproxyd/Virtproxyd_config[log_filters]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54535]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-remote-probe-interval=60000 Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtproxyd/Virtproxyd_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-remote-probe-interval]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtqemud/Virtqemud_config[log_filters]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtqemud/Virtqemud_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtnodedevd/Virtnodedevd_config[log_filters]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtnodedevd/Virtnodedevd_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54537]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-openflow-probe-interval=60 Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtstoraged/Virtstoraged_config[log_filters]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-openflow-probe-interval]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtstoraged/Virtstoraged_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtsecretd/Virtsecretd_config[log_filters]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Virtsecretd/Virtsecretd_config[log_outputs]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtnodedevd_config[unix_sock_group]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtnodedevd_config[auth_unix_ro]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtnodedevd_config[auth_unix_rw]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54541]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-monitor-all=true Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-monitor-all]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtnodedevd_config[unix_sock_ro_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtnodedevd_config[unix_sock_rw_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtproxyd_config[unix_sock_group]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtproxyd_config[auth_unix_ro]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtproxyd_config[auth_unix_rw]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtproxyd_config[unix_sock_ro_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtproxyd_config[unix_sock_rw_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtqemud_config[unix_sock_group]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtqemud_config[auth_unix_ro]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtqemud_config[auth_unix_rw]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54544]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-ofctrl-wait-before-clear=8000 Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtqemud_config[unix_sock_ro_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtqemud_config[unix_sock_rw_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-ofctrl-wait-before-clear]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtsecretd_config[unix_sock_group]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtsecretd_config[auth_unix_ro]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtsecretd_config[auth_unix_rw]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtsecretd_config[unix_sock_ro_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtsecretd_config[unix_sock_rw_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtstoraged_config[unix_sock_group]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtstoraged_config[auth_unix_ro]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtstoraged_config[auth_unix_rw]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtstoraged_config[unix_sock_ro_perms]/ensure: created Nov 27 03:01:04 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Config/Virtstoraged_config[unix_sock_rw_perms]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54546]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-encap-tos=0 Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-encap-tos]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54548]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-chassis-mac-mappings=datacentre:fa:16:3e:5a:d4:f8 Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-chassis-mac-mappings]/ensure: created Nov 27 03:01:04 localhost podman[53353]: 2025-11-27 08:00:58.205363667 +0000 UTC m=+0.044001455 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1 Nov 27 03:01:04 localhost ovs-vsctl[54562]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-bridge-mappings=datacentre:br-ex Nov 27 03:01:04 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-bridge-mappings]/ensure: created Nov 27 03:01:04 localhost ovs-vsctl[54577]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:ovn-match-northd-version=false Nov 27 03:01:05 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:ovn-match-northd-version]/ensure: created Nov 27 03:01:05 localhost systemd[1]: libpod-fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586.scope: Deactivated successfully. Nov 27 03:01:05 localhost systemd[1]: libpod-fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586.scope: Consumed 2.502s CPU time. Nov 27 03:01:05 localhost ovs-vsctl[54584]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl set Open_vSwitch . external_ids:garp-max-timeout-sec=0 Nov 27 03:01:05 localhost puppet-user[54320]: Notice: /Stage[main]/Ovn::Controller/Vs_config[external_ids:garp-max-timeout-sec]/ensure: created Nov 27 03:01:05 localhost puppet-user[54320]: Notice: Applied catalog in 0.47 seconds Nov 27 03:01:05 localhost puppet-user[54320]: Application: Nov 27 03:01:05 localhost puppet-user[54320]: Initial environment: production Nov 27 03:01:05 localhost puppet-user[54320]: Converged environment: production Nov 27 03:01:05 localhost puppet-user[54320]: Run mode: user Nov 27 03:01:05 localhost puppet-user[54320]: Changes: Nov 27 03:01:05 localhost puppet-user[54320]: Total: 14 Nov 27 03:01:05 localhost puppet-user[54320]: Events: Nov 27 03:01:05 localhost puppet-user[54320]: Success: 14 Nov 27 03:01:05 localhost puppet-user[54320]: Total: 14 Nov 27 03:01:05 localhost puppet-user[54320]: Resources: Nov 27 03:01:05 localhost puppet-user[54320]: Skipped: 12 Nov 27 03:01:05 localhost puppet-user[54320]: Changed: 14 Nov 27 03:01:05 localhost puppet-user[54320]: Out of sync: 14 Nov 27 03:01:05 localhost puppet-user[54320]: Total: 29 Nov 27 03:01:05 localhost puppet-user[54320]: Time: Nov 27 03:01:05 localhost puppet-user[54320]: Exec: 0.02 Nov 27 03:01:05 localhost puppet-user[54320]: Config retrieval: 0.29 Nov 27 03:01:05 localhost puppet-user[54320]: Vs config: 0.38 Nov 27 03:01:05 localhost puppet-user[54320]: Transaction evaluation: 0.46 Nov 27 03:01:05 localhost puppet-user[54320]: Catalog application: 0.47 Nov 27 03:01:05 localhost puppet-user[54320]: Last run: 1764230465 Nov 27 03:01:05 localhost puppet-user[54320]: Total: 0.47 Nov 27 03:01:05 localhost puppet-user[54320]: Version: Nov 27 03:01:05 localhost puppet-user[54320]: Config: 1764230464 Nov 27 03:01:05 localhost puppet-user[54320]: Puppet: 7.10.0 Nov 27 03:01:05 localhost podman[54608]: 2025-11-27 08:01:05.104058886 +0000 UTC m=+0.052772340 container died fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, architecture=x86_64, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, vendor=Red Hat, Inc., config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, name=rhosp17/openstack-rsyslog, com.redhat.component=openstack-rsyslog-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_puppet_step1, release=1761123044, tcib_managed=true, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, build-date=2025-11-18T22:49:49Z, managed_by=tripleo_ansible, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, container_name=container-puppet-rsyslog, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:01:05 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:05 localhost systemd[1]: var-lib-containers-storage-overlay-0034e3c51dba6c91c611dd6ead3f9e70360969b5749ffdec598d74f5bdf45edb-merged.mount: Deactivated successfully. Nov 27 03:01:05 localhost podman[54608]: 2025-11-27 08:01:05.267163716 +0000 UTC m=+0.215877130 container cleanup fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=container-puppet-rsyslog, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:49Z, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, config_id=tripleo_puppet_step1, vcs-type=git, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, com.redhat.component=openstack-rsyslog-container, container_name=container-puppet-rsyslog, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.buildah.version=1.41.4, name=rhosp17/openstack-rsyslog) Nov 27 03:01:05 localhost systemd[1]: libpod-conmon-fd3638e2e350062cb79a549dbd9f887e37e9bee7edf04b546757d57a0bf43586.scope: Deactivated successfully. Nov 27 03:01:05 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-rsyslog --conmon-pidfile /run/container-puppet-rsyslog.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment --env NAME=rsyslog --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::logging::rsyslog --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-rsyslog --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,rsyslog::generate_concat,concat::fragment', 'NAME': 'rsyslog', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::logging::rsyslog'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-rsyslog.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 Nov 27 03:01:05 localhost podman[54627]: 2025-11-27 08:01:05.308233807 +0000 UTC m=+0.223869927 container create f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, architecture=x86_64, release=1761123044, batch=17.1_20251118.1, io.openshift.expose-services=, version=17.1.12, build-date=2025-11-19T00:11:59Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, url=https://www.redhat.com, vcs-type=git, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_id=tripleo_puppet_step1, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-central, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, description=Red Hat OpenStack Platform 17.1 ceilometer-central, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-ceilometer, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-central-container) Nov 27 03:01:05 localhost podman[54627]: 2025-11-27 08:01:05.128394116 +0000 UTC m=+0.044030236 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1 Nov 27 03:01:05 localhost systemd[1]: Started libpod-conmon-f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31.scope. Nov 27 03:01:05 localhost systemd[1]: Started libcrun container. Nov 27 03:01:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/e2514e2bf9551d4e31dfc8e5708aee30a6168a670160c4032d4ad1dda4eb3c51/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:05 localhost podman[54627]: 2025-11-27 08:01:05.381973265 +0000 UTC m=+0.297609365 container init f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, io.buildah.version=1.41.4, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-central, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=container-puppet-ceilometer, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.expose-services=, config_id=tripleo_puppet_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:59Z, managed_by=tripleo_ansible, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, name=rhosp17/openstack-ceilometer-central, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, release=1761123044, com.redhat.component=openstack-ceilometer-central-container, architecture=x86_64) Nov 27 03:01:05 localhost podman[54627]: 2025-11-27 08:01:05.395084487 +0000 UTC m=+0.310720607 container start f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_puppet_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-central, maintainer=OpenStack TripleO Team, release=1761123044, tcib_managed=true, com.redhat.component=openstack-ceilometer-central-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-central, url=https://www.redhat.com, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=container-puppet-ceilometer, build-date=2025-11-19T00:11:59Z) Nov 27 03:01:05 localhost podman[54627]: 2025-11-27 08:01:05.395344643 +0000 UTC m=+0.310980763 container attach f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, maintainer=OpenStack TripleO Team, container_name=container-puppet-ceilometer, description=Red Hat OpenStack Platform 17.1 ceilometer-central, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, com.redhat.component=openstack-ceilometer-central-container, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, distribution-scope=public, name=rhosp17/openstack-ceilometer-central, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vcs-type=git, config_id=tripleo_puppet_step1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, tcib_managed=true, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:59Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:01:05 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Compute::Libvirt::Qemu/Augeas[qemu-conf-limits]/returns: executed successfully Nov 27 03:01:05 localhost systemd[1]: libpod-2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b.scope: Deactivated successfully. Nov 27 03:01:05 localhost systemd[1]: libpod-2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b.scope: Consumed 2.908s CPU time. Nov 27 03:01:05 localhost podman[54731]: 2025-11-27 08:01:05.711674197 +0000 UTC m=+0.054072993 container died 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, name=rhosp17/openstack-ovn-controller, vcs-type=git, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vendor=Red Hat, Inc., container_name=container-puppet-ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, batch=17.1_20251118.1, config_id=tripleo_puppet_step1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:01:05 localhost podman[54731]: 2025-11-27 08:01:05.7527555 +0000 UTC m=+0.095154286 container cleanup 2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=container-puppet-ovn_controller, container_name=container-puppet-ovn_controller, version=17.1.12, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, distribution-scope=public, config_id=tripleo_puppet_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-type=git, tcib_managed=true, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, build-date=2025-11-18T23:34:05Z, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:01:05 localhost systemd[1]: libpod-conmon-2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b.scope: Deactivated successfully. Nov 27 03:01:05 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-ovn_controller --conmon-pidfile /run/container-puppet-ovn_controller.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,vs_config,exec --env NAME=ovn_controller --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::neutron::agents::ovn#012 --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-ovn_controller --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,vs_config,exec', 'NAME': 'ovn_controller', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::agents::ovn\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/etc/sysconfig/modules:/etc/sysconfig/modules', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-ovn_controller.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /etc/sysconfig/modules:/etc/sysconfig/modules --volume /lib/modules:/lib/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:01:05 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Migration::Qemu/Augeas[qemu-conf-migration-ports]/returns: executed successfully Nov 27 03:01:05 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Logging/Oslo::Log[nova_config]/Nova_config[DEFAULT/debug]/ensure: created Nov 27 03:01:05 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Logging/Oslo::Log[nova_config]/Nova_config[DEFAULT/log_dir]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cache/Oslo::Cache[nova_config]/Nova_config[cache/backend]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cache/Oslo::Cache[nova_config]/Nova_config[cache/enabled]/ensure: created Nov 27 03:01:06 localhost systemd[1]: var-lib-containers-storage-overlay-973a1ad437b898f3b247f823900f22fbefb6e9a3e82863653f54b4dbb77d854c-merged.mount: Deactivated successfully. Nov 27 03:01:06 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2cce8acd88eafe25ab400a35f9ce24611903ef1851e73b494adffee80377259b-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cache/Oslo::Cache[nova_config]/Nova_config[cache/memcache_servers]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Cache/Oslo::Cache[nova_config]/Nova_config[cache/tls_enabled]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Rabbit[nova_config]/Nova_config[oslo_messaging_rabbit/heartbeat_in_pthread]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Rabbit[nova_config]/Nova_config[oslo_messaging_rabbit/heartbeat_timeout_threshold]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Rabbit[nova_config]/Nova_config[oslo_messaging_rabbit/ssl]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Default[nova_config]/Nova_config[DEFAULT/transport_url]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Notifications[nova_config]/Nova_config[oslo_messaging_notifications/driver]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Messaging::Notifications[nova_config]/Nova_config[oslo_messaging_notifications/transport_url]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova/Oslo::Concurrency[nova_config]/Nova_config[oslo_concurrency/lock_path]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/auth_type]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/region_name]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/auth_url]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/username]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/password]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/user_domain_name]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/project_name]/ensure: created Nov 27 03:01:06 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/project_domain_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[53564]: Notice: /Stage[main]/Nova::Keystone::Service_user/Keystone::Resource::Service_user[nova_config]/Nova_config[service_user/send_service_user_token]/ensure: created Nov 27 03:01:07 localhost podman[54368]: 2025-11-27 08:01:02.822376918 +0000 UTC m=+0.045978544 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1 Nov 27 03:01:07 localhost puppet-user[53564]: Notice: /Stage[main]/Ssh::Server::Config/Concat[/etc/ssh/sshd_config]/File[/etc/ssh/sshd_config]/ensure: defined content as '{sha256}66a7ab6cc1a19ea5002a5aaa2cfb2f196778c89c859d0afac926fe3fac9c75a4' Nov 27 03:01:07 localhost puppet-user[53564]: Notice: Applied catalog in 4.42 seconds Nov 27 03:01:07 localhost puppet-user[53564]: Application: Nov 27 03:01:07 localhost puppet-user[53564]: Initial environment: production Nov 27 03:01:07 localhost puppet-user[53564]: Converged environment: production Nov 27 03:01:07 localhost puppet-user[53564]: Run mode: user Nov 27 03:01:07 localhost puppet-user[53564]: Changes: Nov 27 03:01:07 localhost puppet-user[53564]: Total: 183 Nov 27 03:01:07 localhost puppet-user[53564]: Events: Nov 27 03:01:07 localhost puppet-user[53564]: Success: 183 Nov 27 03:01:07 localhost puppet-user[53564]: Total: 183 Nov 27 03:01:07 localhost puppet-user[53564]: Resources: Nov 27 03:01:07 localhost puppet-user[53564]: Changed: 183 Nov 27 03:01:07 localhost puppet-user[53564]: Out of sync: 183 Nov 27 03:01:07 localhost puppet-user[53564]: Skipped: 57 Nov 27 03:01:07 localhost puppet-user[53564]: Total: 487 Nov 27 03:01:07 localhost puppet-user[53564]: Time: Nov 27 03:01:07 localhost puppet-user[53564]: Concat fragment: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: Concat file: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: Anchor: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: File line: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: Virtlogd config: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: Virtstoraged config: 0.01 Nov 27 03:01:07 localhost puppet-user[53564]: Virtqemud config: 0.01 Nov 27 03:01:07 localhost puppet-user[53564]: Exec: 0.01 Nov 27 03:01:07 localhost puppet-user[53564]: Virtsecretd config: 0.02 Nov 27 03:01:07 localhost puppet-user[53564]: Virtnodedevd config: 0.03 Nov 27 03:01:07 localhost puppet-user[53564]: Virtproxyd config: 0.03 Nov 27 03:01:07 localhost puppet-user[53564]: File: 0.03 Nov 27 03:01:07 localhost puppet-user[53564]: Package: 0.03 Nov 27 03:01:07 localhost puppet-user[53564]: Augeas: 0.98 Nov 27 03:01:07 localhost puppet-user[53564]: Config retrieval: 1.58 Nov 27 03:01:07 localhost puppet-user[53564]: Last run: 1764230467 Nov 27 03:01:07 localhost puppet-user[53564]: Nova config: 3.03 Nov 27 03:01:07 localhost puppet-user[53564]: Transaction evaluation: 4.41 Nov 27 03:01:07 localhost puppet-user[53564]: Catalog application: 4.42 Nov 27 03:01:07 localhost puppet-user[53564]: Resources: 0.00 Nov 27 03:01:07 localhost puppet-user[53564]: Total: 4.42 Nov 27 03:01:07 localhost puppet-user[53564]: Version: Nov 27 03:01:07 localhost puppet-user[53564]: Config: 1764230461 Nov 27 03:01:07 localhost puppet-user[53564]: Puppet: 7.10.0 Nov 27 03:01:07 localhost podman[54808]: 2025-11-27 08:01:07.300607739 +0000 UTC m=+0.102181170 container create e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, vcs-type=git, maintainer=OpenStack TripleO Team, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, description=Red Hat OpenStack Platform 17.1 neutron-server, container_name=container-puppet-neutron, summary=Red Hat OpenStack Platform 17.1 neutron-server, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server, config_id=tripleo_puppet_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, batch=17.1_20251118.1, build-date=2025-11-19T00:23:27Z, com.redhat.component=openstack-neutron-server-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vendor=Red Hat, Inc., tcib_managed=true, release=1761123044, name=rhosp17/openstack-neutron-server, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com) Nov 27 03:01:07 localhost systemd[1]: Started libpod-conmon-e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8.scope. Nov 27 03:01:07 localhost podman[54808]: 2025-11-27 08:01:07.249888738 +0000 UTC m=+0.051462199 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1 Nov 27 03:01:07 localhost systemd[1]: Started libcrun container. Nov 27 03:01:07 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/417f2e5776f4f3986588cc4cdbebd99c8fa2ed452d79909982094d55066039b4/merged/var/lib/config-data supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:07 localhost podman[54808]: 2025-11-27 08:01:07.366805149 +0000 UTC m=+0.168378560 container init e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-server, release=1761123044, io.openshift.expose-services=, vcs-type=git, batch=17.1_20251118.1, version=17.1.12, build-date=2025-11-19T00:23:27Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, name=rhosp17/openstack-neutron-server, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-server-container, io.buildah.version=1.41.4, tcib_managed=true, url=https://www.redhat.com, container_name=container-puppet-neutron, config_id=tripleo_puppet_step1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-server, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:01:07 localhost podman[54808]: 2025-11-27 08:01:07.376949429 +0000 UTC m=+0.178522860 container start e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, version=17.1.12, name=rhosp17/openstack-neutron-server, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server, io.buildah.version=1.41.4, build-date=2025-11-19T00:23:27Z, com.redhat.component=openstack-neutron-server-container, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, batch=17.1_20251118.1, container_name=container-puppet-neutron, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-server, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_id=tripleo_puppet_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-server, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, tcib_managed=true) Nov 27 03:01:07 localhost podman[54808]: 2025-11-27 08:01:07.377515433 +0000 UTC m=+0.179088874 container attach e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-server, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, build-date=2025-11-19T00:23:27Z, distribution-scope=public, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, summary=Red Hat OpenStack Platform 17.1 neutron-server, name=rhosp17/openstack-neutron-server, maintainer=OpenStack TripleO Team, config_id=tripleo_puppet_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, version=17.1.12, tcib_managed=true, vendor=Red Hat, Inc., container_name=container-puppet-neutron, com.redhat.component=openstack-neutron-server-container, vcs-type=git, managed_by=tripleo_ansible, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server, architecture=x86_64) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:07 localhost puppet-user[54705]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:07 localhost puppet-user[54705]: (file & line not available) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:07 localhost puppet-user[54705]: (file & line not available) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_backend'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 145, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::memcache_servers'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 146, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_tls_enabled'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 147, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_tls_cafile'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 148, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_tls_certfile'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 149, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_tls_keyfile'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 150, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::cache_tls_allowed_ciphers'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 151, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::manage_backend_package'. (file: /etc/puppet/modules/ceilometer/manifests/cache.pp, line: 152, column: 39) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_password'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 63, column: 25) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_url'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 68, column: 25) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_region'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 69, column: 28) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_user'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 70, column: 25) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_tenant_name'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 71, column: 29) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_cacert'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 72, column: 23) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_endpoint_type'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 73, column: 26) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_user_domain_name'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 74, column: 33) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_project_domain_name'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 75, column: 36) Nov 27 03:01:07 localhost puppet-user[54705]: Warning: Unknown variable: '::ceilometer::agent::auth::auth_type'. (file: /etc/puppet/modules/ceilometer/manifests/agent/service_credentials.pp, line: 76, column: 26) Nov 27 03:01:07 localhost puppet-user[54705]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.39 seconds Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Ceilometer_config[DEFAULT/http_timeout]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Ceilometer_config[DEFAULT/host]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Ceilometer_config[publisher/telemetry_secret]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Ceilometer_config[hardware/readonly_user_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Ceilometer_config[hardware/readonly_user_password]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/auth_url]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/region_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/username]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/password]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/project_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/interface]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/user_domain_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/project_domain_name]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Service_credentials/Ceilometer_config[service_credentials/auth_type]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Polling/Ceilometer_config[compute/instance_discovery_method]/ensure: created Nov 27 03:01:07 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Polling/Ceilometer_config[DEFAULT/polling_namespaces]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Polling/Ceilometer_config[polling/tenant_name_discovery]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Agent::Polling/Ceilometer_config[coordination/backend_url]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Cache/Oslo::Cache[ceilometer_config]/Ceilometer_config[cache/backend]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Cache/Oslo::Cache[ceilometer_config]/Ceilometer_config[cache/enabled]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Cache/Oslo::Cache[ceilometer_config]/Ceilometer_config[cache/memcache_servers]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Cache/Oslo::Cache[ceilometer_config]/Ceilometer_config[cache/tls_enabled]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Rabbit[ceilometer_config]/Ceilometer_config[oslo_messaging_rabbit/heartbeat_in_pthread]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Amqp[ceilometer_config]/Ceilometer_config[oslo_messaging_amqp/rpc_address_prefix]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Amqp[ceilometer_config]/Ceilometer_config[oslo_messaging_amqp/notify_address_prefix]/ensure: created Nov 27 03:01:08 localhost systemd[1]: libpod-20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b.scope: Deactivated successfully. Nov 27 03:01:08 localhost systemd[1]: libpod-20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b.scope: Consumed 8.739s CPU time. Nov 27 03:01:08 localhost podman[53447]: 2025-11-27 08:01:08.183608475 +0000 UTC m=+9.913994915 container died 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-libvirt-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, vcs-type=git, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-libvirt, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:35:22Z, url=https://www.redhat.com, release=1761123044, version=17.1.12, config_id=tripleo_puppet_step1, container_name=container-puppet-nova_libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt) Nov 27 03:01:08 localhost systemd[1]: tmp-crun.1MKtNl.mount: Deactivated successfully. Nov 27 03:01:08 localhost systemd[1]: var-lib-containers-storage-overlay-55e8351b1958513e36671035d0bb47863b4e87c80590aade3f4e58207a1d6315-merged.mount: Deactivated successfully. Nov 27 03:01:08 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Notifications[ceilometer_config]/Ceilometer_config[oslo_messaging_notifications/driver]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Notifications[ceilometer_config]/Ceilometer_config[oslo_messaging_notifications/transport_url]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Notifications[ceilometer_config]/Ceilometer_config[oslo_messaging_notifications/topics]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer/Oslo::Messaging::Default[ceilometer_config]/Ceilometer_config[DEFAULT/transport_url]/ensure: created Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Logging/Oslo::Log[ceilometer_config]/Ceilometer_config[DEFAULT/debug]/ensure: created Nov 27 03:01:08 localhost podman[54938]: 2025-11-27 08:01:08.335387434 +0000 UTC m=+0.139394695 container cleanup 20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=container-puppet-nova_libvirt, com.redhat.component=openstack-nova-libvirt-container, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:35:22Z, maintainer=OpenStack TripleO Team, container_name=container-puppet-nova_libvirt, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_puppet_step1, managed_by=tripleo_ansible, url=https://www.redhat.com, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, name=rhosp17/openstack-nova-libvirt, release=1761123044, vendor=Red Hat, Inc.) Nov 27 03:01:08 localhost systemd[1]: libpod-conmon-20e3567ef689d774347c20824935d016f0daf343ac3f0487f05b46fcac62d10b.scope: Deactivated successfully. Nov 27 03:01:08 localhost puppet-user[54705]: Notice: /Stage[main]/Ceilometer::Logging/Oslo::Log[ceilometer_config]/Ceilometer_config[DEFAULT/log_dir]/ensure: created Nov 27 03:01:08 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-nova_libvirt --conmon-pidfile /run/container-puppet-nova_libvirt.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password --env NAME=nova_libvirt --env STEP_CONFIG=include ::tripleo::packages#012# TODO(emilien): figure how to deal with libvirt profile.#012# We'll probably treat it like we do with Neutron plugins.#012# Until then, just include it in the default nova-compute role.#012include tripleo::profile::base::nova::compute::libvirt#012#012include tripleo::profile::base::nova::libvirt#012#012include tripleo::profile::base::nova::compute::libvirt_guests#012#012include tripleo::profile::base::sshd#012include tripleo::profile::base::nova::migration::target --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-nova_libvirt --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,nova_config,libvirtd_config,virtlogd_config,virtproxyd_config,virtqemud_config,virtnodedevd_config,virtsecretd_config,virtstoraged_config,nova_config,file,libvirt_tls_password,libvirtd_config,nova_config,file,libvirt_tls_password', 'NAME': 'nova_libvirt', 'STEP_CONFIG': "include ::tripleo::packages\n# TODO(emilien): figure how to deal with libvirt profile.\n# We'll probably treat it like we do with Neutron plugins.\n# Until then, just include it in the default nova-compute role.\ninclude tripleo::profile::base::nova::compute::libvirt\n\ninclude tripleo::profile::base::nova::libvirt\n\ninclude tripleo::profile::base::nova::compute::libvirt_guests\n\ninclude tripleo::profile::base::sshd\ninclude tripleo::profile::base::nova::migration::target"}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-nova_libvirt.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:01:08 localhost puppet-user[54705]: Notice: Applied catalog in 0.50 seconds Nov 27 03:01:08 localhost puppet-user[54705]: Application: Nov 27 03:01:08 localhost puppet-user[54705]: Initial environment: production Nov 27 03:01:08 localhost puppet-user[54705]: Converged environment: production Nov 27 03:01:08 localhost puppet-user[54705]: Run mode: user Nov 27 03:01:08 localhost puppet-user[54705]: Changes: Nov 27 03:01:08 localhost puppet-user[54705]: Total: 31 Nov 27 03:01:08 localhost puppet-user[54705]: Events: Nov 27 03:01:08 localhost puppet-user[54705]: Success: 31 Nov 27 03:01:08 localhost puppet-user[54705]: Total: 31 Nov 27 03:01:08 localhost puppet-user[54705]: Resources: Nov 27 03:01:08 localhost puppet-user[54705]: Skipped: 22 Nov 27 03:01:08 localhost puppet-user[54705]: Changed: 31 Nov 27 03:01:08 localhost puppet-user[54705]: Out of sync: 31 Nov 27 03:01:08 localhost puppet-user[54705]: Total: 151 Nov 27 03:01:08 localhost puppet-user[54705]: Time: Nov 27 03:01:08 localhost puppet-user[54705]: Package: 0.03 Nov 27 03:01:08 localhost puppet-user[54705]: Ceilometer config: 0.41 Nov 27 03:01:08 localhost puppet-user[54705]: Config retrieval: 0.47 Nov 27 03:01:08 localhost puppet-user[54705]: Transaction evaluation: 0.49 Nov 27 03:01:08 localhost puppet-user[54705]: Catalog application: 0.50 Nov 27 03:01:08 localhost puppet-user[54705]: Last run: 1764230468 Nov 27 03:01:08 localhost puppet-user[54705]: Resources: 0.00 Nov 27 03:01:08 localhost puppet-user[54705]: Total: 0.50 Nov 27 03:01:08 localhost puppet-user[54705]: Version: Nov 27 03:01:08 localhost puppet-user[54705]: Config: 1764230467 Nov 27 03:01:08 localhost puppet-user[54705]: Puppet: 7.10.0 Nov 27 03:01:08 localhost systemd[1]: libpod-f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31.scope: Deactivated successfully. Nov 27 03:01:08 localhost systemd[1]: libpod-f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31.scope: Consumed 3.186s CPU time. Nov 27 03:01:08 localhost podman[54627]: 2025-11-27 08:01:08.802002482 +0000 UTC m=+3.717638622 container died f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, vcs-type=git, architecture=x86_64, container_name=container-puppet-ceilometer, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, com.redhat.component=openstack-ceilometer-central-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, description=Red Hat OpenStack Platform 17.1 ceilometer-central, io.openshift.expose-services=, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, tcib_managed=true, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:11:59Z, config_id=tripleo_puppet_step1, name=rhosp17/openstack-ceilometer-central, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:01:08 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:08 localhost systemd[1]: var-lib-containers-storage-overlay-e2514e2bf9551d4e31dfc8e5708aee30a6168a670160c4032d4ad1dda4eb3c51-merged.mount: Deactivated successfully. Nov 27 03:01:08 localhost podman[55011]: 2025-11-27 08:01:08.929402061 +0000 UTC m=+0.116096132 container cleanup f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1, name=container-puppet-ceilometer, config_id=tripleo_puppet_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-central, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, description=Red Hat OpenStack Platform 17.1 ceilometer-central, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-central, com.redhat.component=openstack-ceilometer-central-container, vendor=Red Hat, Inc., io.openshift.expose-services=, batch=17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-central, name=rhosp17/openstack-ceilometer-central, architecture=x86_64, release=1761123044, container_name=container-puppet-ceilometer, build-date=2025-11-19T00:11:59Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-central, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, vcs-type=git, url=https://www.redhat.com, managed_by=tripleo_ansible) Nov 27 03:01:08 localhost systemd[1]: libpod-conmon-f2b09eff8f2430edd47bc1d45b8906072ac9ee09c363cf17dc4869de31ed4a31.scope: Deactivated successfully. Nov 27 03:01:08 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-ceilometer --conmon-pidfile /run/container-puppet-ceilometer.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config --env NAME=ceilometer --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::ceilometer::agent::polling#012include tripleo::profile::base::ceilometer::agent::polling#012 --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-ceilometer --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,ceilometer_config,ceilometer_config', 'NAME': 'ceilometer', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::ceilometer::agent::polling\ninclude tripleo::profile::base::ceilometer::agent::polling\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-ceilometer.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-ceilometer-central:17.1 Nov 27 03:01:09 localhost puppet-user[54910]: Error: Facter: error while resolving custom fact "haproxy_version": undefined method `strip' for nil:NilClass Nov 27 03:01:09 localhost puppet-user[54910]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:01:09 localhost puppet-user[54910]: (file: /etc/puppet/hiera.yaml) Nov 27 03:01:09 localhost puppet-user[54910]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:01:09 localhost puppet-user[54910]: (file & line not available) Nov 27 03:01:09 localhost puppet-user[54910]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:01:09 localhost puppet-user[54910]: (file & line not available) Nov 27 03:01:09 localhost puppet-user[54910]: Warning: Unknown variable: 'dhcp_agents_per_net'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/neutron.pp, line: 154, column: 37) Nov 27 03:01:10 localhost puppet-user[54910]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.73 seconds Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/auth_strategy]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/core_plugin]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/host]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/dns_domain]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/dhcp_agent_notification]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/allow_overlapping_ips]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/global_physnet_mtu]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/vlan_transparent]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[agent/root_helper]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[agent/report_interval]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Neutron_config[DEFAULT/service_plugins]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/debug]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/nova_metadata_host]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/nova_metadata_protocol]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/metadata_proxy_shared_secret]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/metadata_workers]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/state_path]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[DEFAULT/hwol_qos_enabled]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[agent/root_helper]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[ovs/ovsdb_connection]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[ovs/ovsdb_connection_timeout]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[ovn/ovsdb_probe_interval]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[ovn/ovn_nb_connection]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Agents::Ovn_metadata/Ovn_metadata_agent_config[ovn/ovn_sb_connection]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Default[neutron_config]/Neutron_config[DEFAULT/transport_url]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Default[neutron_config]/Neutron_config[DEFAULT/control_exchange]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Concurrency[neutron_config]/Neutron_config[oslo_concurrency/lock_path]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Notifications[neutron_config]/Neutron_config[oslo_messaging_notifications/driver]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Notifications[neutron_config]/Neutron_config[oslo_messaging_notifications/transport_url]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Rabbit[neutron_config]/Neutron_config[oslo_messaging_rabbit/heartbeat_in_pthread]/ensure: created Nov 27 03:01:10 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron/Oslo::Messaging::Rabbit[neutron_config]/Neutron_config[oslo_messaging_rabbit/heartbeat_timeout_threshold]/ensure: created Nov 27 03:01:11 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Logging/Oslo::Log[neutron_config]/Neutron_config[DEFAULT/debug]/ensure: created Nov 27 03:01:11 localhost puppet-user[54910]: Notice: /Stage[main]/Neutron::Logging/Oslo::Log[neutron_config]/Neutron_config[DEFAULT/log_dir]/ensure: created Nov 27 03:01:11 localhost puppet-user[54910]: Notice: Applied catalog in 0.50 seconds Nov 27 03:01:11 localhost puppet-user[54910]: Application: Nov 27 03:01:11 localhost puppet-user[54910]: Initial environment: production Nov 27 03:01:11 localhost puppet-user[54910]: Converged environment: production Nov 27 03:01:11 localhost puppet-user[54910]: Run mode: user Nov 27 03:01:11 localhost puppet-user[54910]: Changes: Nov 27 03:01:11 localhost puppet-user[54910]: Total: 33 Nov 27 03:01:11 localhost puppet-user[54910]: Events: Nov 27 03:01:11 localhost puppet-user[54910]: Success: 33 Nov 27 03:01:11 localhost puppet-user[54910]: Total: 33 Nov 27 03:01:11 localhost puppet-user[54910]: Resources: Nov 27 03:01:11 localhost puppet-user[54910]: Skipped: 21 Nov 27 03:01:11 localhost puppet-user[54910]: Changed: 33 Nov 27 03:01:11 localhost puppet-user[54910]: Out of sync: 33 Nov 27 03:01:11 localhost puppet-user[54910]: Total: 155 Nov 27 03:01:11 localhost puppet-user[54910]: Time: Nov 27 03:01:11 localhost puppet-user[54910]: Resources: 0.00 Nov 27 03:01:11 localhost puppet-user[54910]: Ovn metadata agent config: 0.02 Nov 27 03:01:11 localhost puppet-user[54910]: Neutron config: 0.40 Nov 27 03:01:11 localhost puppet-user[54910]: Transaction evaluation: 0.49 Nov 27 03:01:11 localhost puppet-user[54910]: Catalog application: 0.50 Nov 27 03:01:11 localhost puppet-user[54910]: Config retrieval: 0.81 Nov 27 03:01:11 localhost puppet-user[54910]: Last run: 1764230471 Nov 27 03:01:11 localhost puppet-user[54910]: Total: 0.50 Nov 27 03:01:11 localhost puppet-user[54910]: Version: Nov 27 03:01:11 localhost puppet-user[54910]: Config: 1764230469 Nov 27 03:01:11 localhost puppet-user[54910]: Puppet: 7.10.0 Nov 27 03:01:11 localhost systemd[1]: libpod-e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8.scope: Deactivated successfully. Nov 27 03:01:11 localhost systemd[1]: libpod-e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8.scope: Consumed 4.009s CPU time. Nov 27 03:01:11 localhost podman[54808]: 2025-11-27 08:01:11.733786959 +0000 UTC m=+4.535360420 container died e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, name=rhosp17/openstack-neutron-server, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, vendor=Red Hat, Inc., url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-server-container, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, release=1761123044, config_id=tripleo_puppet_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-server, summary=Red Hat OpenStack Platform 17.1 neutron-server, build-date=2025-11-19T00:23:27Z, container_name=container-puppet-neutron, batch=17.1_20251118.1) Nov 27 03:01:11 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:11 localhost systemd[1]: var-lib-containers-storage-overlay-417f2e5776f4f3986588cc4cdbebd99c8fa2ed452d79909982094d55066039b4-merged.mount: Deactivated successfully. Nov 27 03:01:11 localhost podman[55159]: 2025-11-27 08:01:11.867582936 +0000 UTC m=+0.121277959 container cleanup e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1, name=container-puppet-neutron, vendor=Red Hat, Inc., release=1761123044, version=17.1.12, container_name=container-puppet-neutron, com.redhat.component=openstack-neutron-server-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-server, config_id=tripleo_puppet_step1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-server, config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']}, url=https://www.redhat.com, batch=17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, build-date=2025-11-19T00:23:27Z, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-server, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-server, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-server, distribution-scope=public, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-server) Nov 27 03:01:11 localhost systemd[1]: libpod-conmon-e07b895a5a83ac8b45b2cbafda69d82aef0c9687561498459884fcb91698adf8.scope: Deactivated successfully. Nov 27 03:01:11 localhost python3[53256]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name container-puppet-neutron --conmon-pidfile /run/container-puppet-neutron.pid --detach=False --entrypoint /var/lib/container-puppet/container-puppet.sh --env STEP=6 --env NET_HOST=true --env DEBUG=true --env HOSTNAME=np0005537446 --env NO_ARCHIVE= --env PUPPET_TAGS=file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config --env NAME=neutron --env STEP_CONFIG=include ::tripleo::packages#012include tripleo::profile::base::neutron::ovn_metadata#012 --label config_id=tripleo_puppet_step1 --label container_name=container-puppet-neutron --label managed_by=tripleo_ansible --label config_data={'security_opt': ['label=disable'], 'user': 0, 'detach': False, 'recreate': True, 'entrypoint': '/var/lib/container-puppet/container-puppet.sh', 'environment': {'STEP': 6, 'NET_HOST': 'true', 'DEBUG': 'true', 'HOSTNAME': 'np0005537446', 'NO_ARCHIVE': '', 'PUPPET_TAGS': 'file,file_line,concat,augeas,cron,neutron_config,ovn_metadata_agent_config', 'NAME': 'neutron', 'STEP_CONFIG': 'include ::tripleo::packages\ninclude tripleo::profile::base::neutron::ovn_metadata\n'}, 'net': ['host'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1', 'volumes': ['/dev/log:/dev/log:rw', '/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/config-data:/var/lib/config-data:rw', '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', '/var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/container-puppet-neutron.log --network host --security-opt label=disable --user 0 --volume /dev/log:/dev/log:rw --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /lib/modules:/lib/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/config-data:/var/lib/config-data:rw --volume /var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro --volume /var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro --volume /var/lib/container-puppet/puppetlabs:/opt/puppetlabs:ro registry.redhat.io/rhosp-rhel9/openstack-neutron-server:17.1 Nov 27 03:01:12 localhost python3[55214]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:13 localhost python3[55246]: ansible-stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:01:14 localhost python3[55296]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-container-shutdown follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:14 localhost python3[55339]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230473.9236195-84173-119504943320582/source dest=/usr/libexec/tripleo-container-shutdown mode=0700 owner=root group=root _original_basename=tripleo-container-shutdown follow=False checksum=7d67b1986212f5548057505748cd74cfcf9c0d35 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:15 localhost python3[55401]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-start-podman-container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:15 localhost python3[55444]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230474.7798498-84173-161430668261793/source dest=/usr/libexec/tripleo-start-podman-container mode=0700 owner=root group=root _original_basename=tripleo-start-podman-container follow=False checksum=536965633b8d3b1ce794269ffb07be0105a560a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:16 localhost python3[55506]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/tripleo-container-shutdown.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:16 localhost python3[55549]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230475.745503-84246-234803741774587/source dest=/usr/lib/systemd/system/tripleo-container-shutdown.service mode=0644 owner=root group=root _original_basename=tripleo-container-shutdown-service follow=False checksum=66c1d41406ba8714feb9ed0a35259a7a57ef9707 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:16 localhost python3[55611]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:17 localhost python3[55654]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230476.6486676-84335-279913401338453/source dest=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset mode=0644 owner=root group=root _original_basename=91-tripleo-container-shutdown-preset follow=False checksum=bccb1207dcbcfaa5ca05f83c8f36ce4c2460f081 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:17 localhost python3[55684]: ansible-systemd Invoked with name=tripleo-container-shutdown state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:01:17 localhost systemd[1]: Reloading. Nov 27 03:01:17 localhost systemd-rc-local-generator[55708]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:17 localhost systemd-sysv-generator[55711]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:18 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:18 localhost systemd[1]: Reloading. Nov 27 03:01:18 localhost systemd-rc-local-generator[55747]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:18 localhost systemd-sysv-generator[55753]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:18 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:18 localhost systemd[1]: Starting TripleO Container Shutdown... Nov 27 03:01:18 localhost systemd[1]: Finished TripleO Container Shutdown. Nov 27 03:01:18 localhost python3[55808]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/netns-placeholder.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:19 localhost python3[55851]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230478.6528645-84373-135392123914260/source dest=/usr/lib/systemd/system/netns-placeholder.service mode=0644 owner=root group=root _original_basename=netns-placeholder-service follow=False checksum=8e9c6d5ce3a6e7f71c18780ec899f32f23de4c71 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:19 localhost python3[55913]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:01:20 localhost python3[55956]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230479.6180496-84387-101087592869651/source dest=/usr/lib/systemd/system-preset/91-netns-placeholder.preset mode=0644 owner=root group=root _original_basename=91-netns-placeholder-preset follow=False checksum=28b7b9aa893525d134a1eeda8a0a48fb25b736b9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:20 localhost python3[55986]: ansible-systemd Invoked with name=netns-placeholder state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:01:20 localhost systemd[1]: Reloading. Nov 27 03:01:20 localhost systemd-rc-local-generator[56009]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:20 localhost systemd-sysv-generator[56013]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:20 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:21 localhost systemd[1]: Reloading. Nov 27 03:01:21 localhost systemd-sysv-generator[56051]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:21 localhost systemd-rc-local-generator[56048]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:21 localhost systemd[1]: Starting Create netns directory... Nov 27 03:01:21 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 03:01:21 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 03:01:21 localhost systemd[1]: Finished Create netns directory. Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for metrics_qdr, new hash: fcb31553ac96170fad0e8a9767bee027 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for collectd, new hash: d31718fcd17fdeee6489534105191c7a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for iscsid, new hash: 2c87de3317f94758d1bec36af3e86047 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtlogd_wrapper, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtnodedevd, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtproxyd, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtqemud, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtsecretd, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_virtstoraged, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for rsyslog, new hash: fd416cb72dcd90fffe349badf5624773 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for ceilometer_agent_compute, new hash: d7f975e76c3815f85351b6258d097495 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for ceilometer_agent_ipmi, new hash: d7f975e76c3815f85351b6258d097495 Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for logrotate_crond, new hash: 53ed83bb0cae779ff95edb2002262c6f Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_libvirt_init_secret, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_migration_target, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for ovn_metadata_agent, new hash: 8cac2334de855b5d814608ea6483099b Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_compute, new hash: 2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:21 localhost python3[56079]: ansible-container_puppet_config [WARNING] Config change detected for nova_wait_for_compute_service, new hash: 4018231ef757a0bfd34181dc6dc2933a Nov 27 03:01:23 localhost python3[56136]: ansible-tripleo_container_manage Invoked with config_id=tripleo_step1 config_dir=/var/lib/tripleo-config/container-startup-config/step_1 config_patterns=*.json config_overrides={} concurrency=5 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:01:23 localhost podman[56174]: 2025-11-27 08:01:23.972383569 +0000 UTC m=+0.087748572 container create 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.openshift.expose-services=, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, container_name=metrics_qdr_init_logs, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:01:24 localhost systemd[1]: Started libpod-conmon-517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89.scope. Nov 27 03:01:24 localhost podman[56174]: 2025-11-27 08:01:23.932847825 +0000 UTC m=+0.048212868 image pull registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 03:01:24 localhost systemd[1]: Started libcrun container. Nov 27 03:01:24 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/494c86aa6aa07ca021d65ef2e92ce055836a68c271a53c36c37577e44fff39b5/merged/var/log/qdrouterd supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:24 localhost podman[56174]: 2025-11-27 08:01:24.053109689 +0000 UTC m=+0.168474692 container init 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, config_id=tripleo_step1, vcs-type=git, build-date=2025-11-18T22:49:46Z, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, container_name=metrics_qdr_init_logs, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., tcib_managed=true) Nov 27 03:01:24 localhost podman[56174]: 2025-11-27 08:01:24.062482299 +0000 UTC m=+0.177847312 container start 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, config_id=tripleo_step1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, container_name=metrics_qdr_init_logs, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public) Nov 27 03:01:24 localhost podman[56174]: 2025-11-27 08:01:24.062812877 +0000 UTC m=+0.178177930 container attach 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, distribution-scope=public, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, batch=17.1_20251118.1, container_name=metrics_qdr_init_logs, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:01:24 localhost systemd[1]: libpod-517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89.scope: Deactivated successfully. Nov 27 03:01:24 localhost podman[56174]: 2025-11-27 08:01:24.071367498 +0000 UTC m=+0.186732531 container died 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, tcib_managed=true, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, vendor=Red Hat, Inc., container_name=metrics_qdr_init_logs, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, version=17.1.12, vcs-type=git, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-qdrouterd) Nov 27 03:01:24 localhost podman[56193]: 2025-11-27 08:01:24.163235422 +0000 UTC m=+0.077920631 container cleanup 517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr_init_logs, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, build-date=2025-11-18T22:49:46Z, tcib_managed=true, name=rhosp17/openstack-qdrouterd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr_init_logs, maintainer=OpenStack TripleO Team) Nov 27 03:01:24 localhost systemd[1]: libpod-conmon-517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89.scope: Deactivated successfully. Nov 27 03:01:24 localhost python3[56136]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name metrics_qdr_init_logs --conmon-pidfile /run/metrics_qdr_init_logs.pid --detach=False --label config_id=tripleo_step1 --label container_name=metrics_qdr_init_logs --label managed_by=tripleo_ansible --label config_data={'command': ['/bin/bash', '-c', 'chown -R qdrouterd:qdrouterd /var/log/qdrouterd'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'none', 'privileged': False, 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/metrics_qdr_init_logs.log --network none --privileged=False --user root --volume /var/log/containers/metrics_qdr:/var/log/qdrouterd:z registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 /bin/bash -c chown -R qdrouterd:qdrouterd /var/log/qdrouterd Nov 27 03:01:24 localhost podman[56271]: 2025-11-27 08:01:24.664563584 +0000 UTC m=+0.085080177 container create 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, container_name=metrics_qdr, vendor=Red Hat, Inc.) Nov 27 03:01:24 localhost systemd[1]: Started libpod-conmon-646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.scope. Nov 27 03:01:24 localhost podman[56271]: 2025-11-27 08:01:24.622268492 +0000 UTC m=+0.042785125 image pull registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 03:01:24 localhost systemd[1]: Started libcrun container. Nov 27 03:01:24 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4e1c0acf5eef7e3d55db68ca8dcde6b2609fb9b33f28ab8549797d29f94ebb77/merged/var/lib/qdrouterd supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:24 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4e1c0acf5eef7e3d55db68ca8dcde6b2609fb9b33f28ab8549797d29f94ebb77/merged/var/log/qdrouterd supports timestamps until 2038 (0x7fffffff) Nov 27 03:01:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:01:24 localhost podman[56271]: 2025-11-27 08:01:24.761718078 +0000 UTC m=+0.182234711 container init 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.buildah.version=1.41.4, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com) Nov 27 03:01:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:01:24 localhost podman[56271]: 2025-11-27 08:01:24.796221888 +0000 UTC m=+0.216738491 container start 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-type=git, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, tcib_managed=true, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_id=tripleo_step1, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team) Nov 27 03:01:24 localhost python3[56136]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name metrics_qdr --conmon-pidfile /run/metrics_qdr.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=fcb31553ac96170fad0e8a9767bee027 --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step1 --label container_name=metrics_qdr --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/metrics_qdr.log --network host --privileged=False --user qdrouterd --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro --volume /var/lib/metrics_qdr:/var/lib/qdrouterd:z --volume /var/log/containers/metrics_qdr:/var/log/qdrouterd:z registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1 Nov 27 03:01:24 localhost podman[56293]: 2025-11-27 08:01:24.904336032 +0000 UTC m=+0.096268443 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=starting, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, batch=17.1_20251118.1, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, distribution-scope=public, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-type=git, architecture=x86_64, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1) Nov 27 03:01:24 localhost systemd[1]: var-lib-containers-storage-overlay-494c86aa6aa07ca021d65ef2e92ce055836a68c271a53c36c37577e44fff39b5-merged.mount: Deactivated successfully. Nov 27 03:01:24 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-517973f8449687ef08a4af288d19d61a68c652120a6d503213a27bb6010d8a89-userdata-shm.mount: Deactivated successfully. Nov 27 03:01:25 localhost podman[56293]: 2025-11-27 08:01:25.129885129 +0000 UTC m=+0.321817580 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-type=git, distribution-scope=public, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044) Nov 27 03:01:25 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:01:25 localhost python3[56367]: ansible-file Invoked with path=/etc/systemd/system/tripleo_metrics_qdr.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:25 localhost python3[56383]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_metrics_qdr_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:01:26 localhost python3[56444]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230485.7712376-84498-23399546210792/source dest=/etc/systemd/system/tripleo_metrics_qdr.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:26 localhost python3[56460]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 03:01:26 localhost systemd[1]: Reloading. Nov 27 03:01:26 localhost systemd-sysv-generator[56486]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:26 localhost systemd-rc-local-generator[56481]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:26 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:27 localhost python3[56512]: ansible-systemd Invoked with state=restarted name=tripleo_metrics_qdr.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:01:27 localhost systemd[1]: Reloading. Nov 27 03:01:27 localhost systemd-rc-local-generator[56537]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:01:27 localhost systemd-sysv-generator[56541]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:01:27 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:01:28 localhost systemd[1]: Starting metrics_qdr container... Nov 27 03:01:28 localhost systemd[1]: Started metrics_qdr container. Nov 27 03:01:28 localhost python3[56592]: ansible-file Invoked with path=/var/lib/container-puppet/container-puppet-tasks1.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:30 localhost python3[56713]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=True puppet_config=/var/lib/container-puppet/container-puppet-tasks1.json short_hostname=np0005537446 step=1 update_config_hash_only=False Nov 27 03:01:30 localhost python3[56729]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:01:30 localhost python3[56745]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_1 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:01:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:01:56 localhost podman[56824]: 2025-11-27 08:01:56.008144596 +0000 UTC m=+0.096648180 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, distribution-scope=public, container_name=metrics_qdr, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:01:56 localhost podman[56824]: 2025-11-27 08:01:56.213227367 +0000 UTC m=+0.301730961 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, tcib_managed=true, vcs-type=git, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, release=1761123044, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, architecture=x86_64) Nov 27 03:01:56 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:02:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:02:26 localhost podman[56853]: 2025-11-27 08:02:26.99454439 +0000 UTC m=+0.088195016 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., architecture=x86_64) Nov 27 03:02:27 localhost podman[56853]: 2025-11-27 08:02:27.198054479 +0000 UTC m=+0.291705145 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, architecture=x86_64, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:02:27 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:02:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:02:57 localhost systemd[1]: tmp-crun.ZsTQXx.mount: Deactivated successfully. Nov 27 03:02:57 localhost podman[56961]: 2025-11-27 08:02:57.995239399 +0000 UTC m=+0.089442759 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.buildah.version=1.41.4, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.openshift.expose-services=, maintainer=OpenStack TripleO Team, config_id=tripleo_step1) Nov 27 03:02:58 localhost podman[56961]: 2025-11-27 08:02:58.223081463 +0000 UTC m=+0.317284843 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, vendor=Red Hat, Inc., container_name=metrics_qdr, version=17.1.12, release=1761123044, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:02:58 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:03:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:03:29 localhost podman[56990]: 2025-11-27 08:03:28.999174701 +0000 UTC m=+0.096979060 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step1, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, vcs-type=git, version=17.1.12, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, build-date=2025-11-18T22:49:46Z, tcib_managed=true) Nov 27 03:03:29 localhost podman[56990]: 2025-11-27 08:03:29.221379505 +0000 UTC m=+0.319183824 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, release=1761123044, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z) Nov 27 03:03:29 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:03:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:03:59 localhost podman[57095]: 2025-11-27 08:03:59.987913688 +0000 UTC m=+0.083429309 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, vcs-type=git, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, tcib_managed=true, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, architecture=x86_64, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}) Nov 27 03:04:00 localhost podman[57095]: 2025-11-27 08:04:00.236203509 +0000 UTC m=+0.331719100 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, architecture=x86_64, config_id=tripleo_step1, distribution-scope=public, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible) Nov 27 03:04:00 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:04:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:04:30 localhost podman[57124]: 2025-11-27 08:04:30.98473239 +0000 UTC m=+0.079171095 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, distribution-scope=public, managed_by=tripleo_ansible, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, batch=17.1_20251118.1, container_name=metrics_qdr) Nov 27 03:04:31 localhost podman[57124]: 2025-11-27 08:04:31.173596883 +0000 UTC m=+0.268035638 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vcs-type=git, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:04:31 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:05:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:05:01 localhost podman[57227]: 2025-11-27 08:05:01.972816222 +0000 UTC m=+0.075101405 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, managed_by=tripleo_ansible, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, architecture=x86_64, io.buildah.version=1.41.4) Nov 27 03:05:02 localhost podman[57227]: 2025-11-27 08:05:02.19010331 +0000 UTC m=+0.292388503 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.buildah.version=1.41.4, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, architecture=x86_64, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, url=https://www.redhat.com, distribution-scope=public, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:05:02 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:05:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:05:32 localhost podman[57257]: 2025-11-27 08:05:32.986959626 +0000 UTC m=+0.081336000 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, version=17.1.12, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 03:05:33 localhost podman[57257]: 2025-11-27 08:05:33.170557019 +0000 UTC m=+0.264933393 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, release=1761123044, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-18T22:49:46Z, tcib_managed=true, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, vcs-type=git) Nov 27 03:05:33 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:06:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:06:03 localhost podman[57363]: 2025-11-27 08:06:03.976706371 +0000 UTC m=+0.076279435 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, batch=17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:06:04 localhost podman[57363]: 2025-11-27 08:06:04.164099286 +0000 UTC m=+0.263672410 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, release=1761123044, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, architecture=x86_64, vendor=Red Hat, Inc., distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, version=17.1.12, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:06:04 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:06:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:06:34 localhost podman[57392]: 2025-11-27 08:06:34.993595515 +0000 UTC m=+0.088022087 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, version=17.1.12, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, architecture=x86_64, io.openshift.expose-services=, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, io.buildah.version=1.41.4) Nov 27 03:06:35 localhost podman[57392]: 2025-11-27 08:06:35.189933357 +0000 UTC m=+0.284359889 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, maintainer=OpenStack TripleO Team, version=17.1.12, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1) Nov 27 03:06:35 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:06:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 17 pg[2.0( empty local-lis/les=0/0 n=0 ec=17/17 lis/c=0/0 les/c/f=0/0/0 sis=17) [3,1,5] r=0 lpr=17 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:06:36 localhost ceph-osd[32722]: osd.3 pg_epoch: 18 pg[2.0( empty local-lis/les=17/18 n=0 ec=17/17 lis/c=0/0 les/c/f=0/0/0 sis=17) [3,1,5] r=0 lpr=17 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:06:38 localhost ceph-osd[32722]: osd.3 pg_epoch: 19 pg[3.0( empty local-lis/les=0/0 n=0 ec=19/19 lis/c=0/0 les/c/f=0/0/0 sis=19) [1,3,2] r=1 lpr=19 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:06:40 localhost ceph-osd[31760]: osd.0 pg_epoch: 21 pg[4.0( empty local-lis/les=0/0 n=0 ec=21/21 lis/c=0/0 les/c/f=0/0/0 sis=21) [5,0,1] r=1 lpr=21 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:06:42 localhost ceph-osd[32722]: osd.3 pg_epoch: 23 pg[5.0( empty local-lis/les=0/0 n=0 ec=23/23 lis/c=0/0 les/c/f=0/0/0 sis=23) [4,5,3] r=2 lpr=23 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:06:58 localhost ceph-osd[31760]: osd.0 pg_epoch: 29 pg[6.0( empty local-lis/les=0/0 n=0 ec=29/29 lis/c=0/0 les/c/f=0/0/0 sis=29) [5,0,1] r=1 lpr=29 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:06:59 localhost ceph-osd[31760]: osd.0 pg_epoch: 31 pg[7.0( empty local-lis/les=0/0 n=0 ec=31/31 lis/c=0/0 les/c/f=0/0/0 sis=31) [0,1,5] r=0 lpr=31 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:00 localhost ceph-osd[31760]: osd.0 pg_epoch: 32 pg[7.0( empty local-lis/les=31/32 n=0 ec=31/31 lis/c=0/0 les/c/f=0/0/0 sis=31) [0,1,5] r=0 lpr=31 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:07:06 localhost podman[57499]: 2025-11-27 08:07:05.999499865 +0000 UTC m=+0.085641622 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.openshift.expose-services=, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:07:06 localhost podman[57499]: 2025-11-27 08:07:06.213953422 +0000 UTC m=+0.300095079 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-type=git, container_name=metrics_qdr, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:46Z, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}) Nov 27 03:07:06 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:07:26 localhost ceph-osd[32722]: osd.3 pg_epoch: 36 pg[3.0( empty local-lis/les=19/20 n=0 ec=19/19 lis/c=19/19 les/c/f=20/20/0 sis=36 pruub=15.952780724s) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 active pruub 1186.177734375s@ mbc={}] start_peering_interval up [1,3,2] -> [1,3,2], acting [1,3,2] -> [1,3,2], acting_primary 1 -> 1, up_primary 1 -> 1, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:26 localhost ceph-osd[32722]: osd.3 pg_epoch: 36 pg[2.0( empty local-lis/les=17/18 n=0 ec=17/17 lis/c=17/17 les/c/f=18/18/0 sis=36 pruub=14.099865913s) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active pruub 1184.324829102s@ mbc={}] start_peering_interval up [3,1,5] -> [3,1,5], acting [3,1,5] -> [3,1,5], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:26 localhost ceph-osd[32722]: osd.3 pg_epoch: 36 pg[3.0( empty local-lis/les=19/20 n=0 ec=19/19 lis/c=19/19 les/c/f=20/20/0 sis=36 pruub=15.949428558s) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1186.177734375s@ mbc={}] state: transitioning to Stray Nov 27 03:07:26 localhost ceph-osd[32722]: osd.3 pg_epoch: 36 pg[2.0( empty local-lis/les=17/18 n=0 ec=17/17 lis/c=17/17 les/c/f=18/18/0 sis=36 pruub=14.099865913s) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown pruub 1184.324829102s@ mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.19( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.17( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.19( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.18( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.18( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.16( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.17( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.15( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.14( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.14( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.13( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.15( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.16( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.12( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.12( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.11( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.13( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.10( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.10( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.11( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.e( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.f( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.d( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.c( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.c( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.d( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.b( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.a( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.a( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.b( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.3( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.2( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.f( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.7( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.6( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.e( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.2( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.3( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.5( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.4( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.5( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.6( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.4( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.8( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.9( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.9( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.8( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1b( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1a( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1a( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1b( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1d( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1c( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1c( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.7( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1d( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1f( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[3.1e( empty local-lis/les=19/20 n=0 ec=36/19 lis/c=19/19 les/c/f=20/20/0 sis=36) [1,3,2] r=1 lpr=36 pi=[19,36)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1e( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1f( empty local-lis/les=17/18 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.0( empty local-lis/les=36/37 n=0 ec=17/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.9( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.6( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.7( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.5( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.4( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.3( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.2( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.19( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.8( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.11( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.10( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.12( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.13( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.15( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.1f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.14( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.18( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.17( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:27 localhost ceph-osd[32722]: osd.3 pg_epoch: 37 pg[2.16( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=17/17 les/c/f=18/18/0 sis=36) [3,1,5] r=0 lpr=36 pi=[17,36)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:28 localhost ceph-osd[32722]: osd.3 pg_epoch: 38 pg[5.0( empty local-lis/les=23/24 n=0 ec=23/23 lis/c=23/23 les/c/f=24/24/0 sis=38 pruub=10.345563889s) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 active pruub 1182.599487305s@ mbc={}] start_peering_interval up [4,5,3] -> [4,5,3], acting [4,5,3] -> [4,5,3], acting_primary 4 -> 4, up_primary 4 -> 4, role 2 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:28 localhost ceph-osd[32722]: osd.3 pg_epoch: 38 pg[5.0( empty local-lis/les=23/24 n=0 ec=23/23 lis/c=23/23 les/c/f=24/24/0 sis=38 pruub=10.343535423s) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1182.599487305s@ mbc={}] state: transitioning to Stray Nov 27 03:07:28 localhost ceph-osd[31760]: osd.0 pg_epoch: 38 pg[4.0( empty local-lis/les=21/22 n=0 ec=21/21 lis/c=21/21 les/c/f=22/22/0 sis=38 pruub=8.412819862s) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 active pruub 1185.146362305s@ mbc={}] start_peering_interval up [5,0,1] -> [5,0,1], acting [5,0,1] -> [5,0,1], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:28 localhost ceph-osd[31760]: osd.0 pg_epoch: 38 pg[4.0( empty local-lis/les=21/22 n=0 ec=21/21 lis/c=21/21 les/c/f=22/22/0 sis=38 pruub=8.410726547s) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1185.146362305s@ mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1e( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.10( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1f( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.13( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.11( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.14( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.15( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.16( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.17( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.8( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.12( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.a( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.b( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.c( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.9( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.d( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.7( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.4( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.6( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.5( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.2( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.3( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.f( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.e( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1d( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1c( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1b( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.19( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1a( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.18( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[32722]: osd.3 pg_epoch: 39 pg[5.1( empty local-lis/les=23/24 n=0 ec=38/23 lis/c=23/23 les/c/f=24/24/0 sis=38) [4,5,3] r=2 lpr=38 pi=[23,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.18( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.14( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.16( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.13( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.12( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.15( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.11( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.17( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.10( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.f( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.e( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.c( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.d( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.b( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.2( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.3( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.19( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.9( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1a( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.5( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.7( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.6( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.a( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.8( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1b( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1c( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.4( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1d( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1e( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:29 localhost ceph-osd[31760]: osd.0 pg_epoch: 39 pg[4.1f( empty local-lis/les=21/22 n=0 ec=38/21 lis/c=21/21 les/c/f=22/22/0 sis=38) [5,0,1] r=1 lpr=38 pi=[21,38)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:30 localhost ceph-osd[31760]: osd.0 pg_epoch: 40 pg[6.0( empty local-lis/les=29/30 n=0 ec=29/29 lis/c=29/29 les/c/f=30/30/0 sis=40 pruub=8.082062721s) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 active pruub 1186.835815430s@ mbc={}] start_peering_interval up [5,0,1] -> [5,0,1], acting [5,0,1] -> [5,0,1], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:30 localhost ceph-osd[31760]: osd.0 pg_epoch: 40 pg[7.0( v 33'39 (0'0,33'39] local-lis/les=31/32 n=22 ec=31/31 lis/c=31/31 les/c/f=32/32/0 sis=40 pruub=10.158322334s) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 33'38 mlcod 33'38 active pruub 1188.913085938s@ mbc={}] start_peering_interval up [0,1,5] -> [0,1,5], acting [0,1,5] -> [0,1,5], acting_primary 0 -> 0, up_primary 0 -> 0, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:30 localhost ceph-osd[31760]: osd.0 pg_epoch: 40 pg[7.0( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=31/31 lis/c=31/31 les/c/f=32/32/0 sis=40 pruub=10.158322334s) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 33'38 mlcod 0'0 unknown pruub 1188.913085938s@ mbc={}] state: transitioning to Primary Nov 27 03:07:30 localhost ceph-osd[31760]: osd.0 pg_epoch: 40 pg[6.0( empty local-lis/les=29/30 n=0 ec=29/29 lis/c=29/29 les/c/f=30/30/0 sis=40 pruub=8.078504562s) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1186.835815430s@ mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1a( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.14( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.17( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.16( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.11( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.13( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.10( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.15( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.12( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.c( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.d( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.c( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.e( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.d( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.e( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.f( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.f( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.9( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.3( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.8( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.1( v 33'39 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.2( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.3( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.2( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1b( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.6( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.7( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.b( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.a( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.6( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.7( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.18( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.8( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.5( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.4( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.4( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.5( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.b( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.a( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1e( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1f( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1c( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.19( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.9( v 33'39 lc 0'0 (0'0,33'39] local-lis/les=31/32 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[6.1d( empty local-lis/les=29/30 n=0 ec=40/29 lis/c=29/29 les/c/f=30/30/0 sis=40) [5,0,1] r=1 lpr=40 pi=[29,40)/1 crt=0'0 mlcod 0'0 unknown NOTIFY mbc={}] state: transitioning to Stray Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.0( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=31/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 33'38 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.1( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.8( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.a( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.c( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.3( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.9( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.d( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.6( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.5( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.2( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.4( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:31 localhost ceph-osd[31760]: osd.0 pg_epoch: 41 pg[7.b( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=31/31 les/c/f=32/32/0 sis=40) [0,1,5] r=0 lpr=40 pi=[31,40)/1 crt=33'39 lcod 0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:32 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.0 scrub starts Nov 27 03:07:32 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.0 scrub ok Nov 27 03:07:33 localhost python3[57574]: ansible-file Invoked with path=/var/lib/tripleo-config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1e( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.890060425s) [5,3,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295532227s@ mbc={}] start_peering_interval up [4,5,3] -> [5,3,4], acting [4,5,3] -> [5,3,4], acting_primary 4 -> 5, up_primary 4 -> 5, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.18( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872400284s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277954102s@ mbc={}] start_peering_interval up [1,3,2] -> [5,1,3], acting [1,3,2] -> [5,1,3], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1f( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.889665604s) [3,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295166016s@ mbc={}] start_peering_interval up [4,5,3] -> [3,4,2], acting [4,5,3] -> [3,4,2], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.19( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872312546s) [2,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [3,1,5] -> [2,0,4], acting [3,1,5] -> [2,0,4], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1e( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.889986992s) [5,3,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295532227s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.19( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872269630s) [2,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277832031s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.18( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872327805s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277954102s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1f( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.889665604s) [3,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.295166016s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.19( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872691154s) [5,0,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278442383s@ mbc={}] start_peering_interval up [1,3,2] -> [5,0,1], acting [1,3,2] -> [5,0,1], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.19( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872672081s) [5,0,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278442383s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.18( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.878116608s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283935547s@ mbc={}] start_peering_interval up [3,1,5] -> [4,3,2], acting [3,1,5] -> [4,3,2], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.17( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.878145218s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.284179688s@ mbc={}] start_peering_interval up [3,1,5] -> [3,1,2], acting [3,1,5] -> [3,1,2], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.10( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.889515877s) [0,4,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295410156s@ mbc={}] start_peering_interval up [4,5,3] -> [0,4,5], acting [4,5,3] -> [0,4,5], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.17( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.878145218s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.284179688s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.18( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.878058434s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283935547s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.10( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.889481544s) [0,4,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295410156s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.16( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872382164s) [3,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278442383s@ mbc={}] start_peering_interval up [1,3,2] -> [3,2,4], acting [1,3,2] -> [3,2,4], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.16( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872382164s) [3,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.278442383s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.11( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.888012886s) [3,4,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294067383s@ mbc={}] start_peering_interval up [4,5,3] -> [3,4,5], acting [4,5,3] -> [3,4,5], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.17( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872149467s) [5,3,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278320312s@ mbc={}] start_peering_interval up [1,3,2] -> [5,3,1], acting [1,3,2] -> [5,3,1], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.11( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.888012886s) [3,4,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.294067383s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.17( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.872132301s) [5,3,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278320312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.12( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887751579s) [1,3,5] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294067383s@ mbc={}] start_peering_interval up [4,5,3] -> [1,3,5], acting [4,5,3] -> [1,3,5], acting_primary 4 -> 1, up_primary 4 -> 1, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.12( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887735367s) [1,3,5] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.294067383s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.15( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877325058s) [1,2,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283691406s@ mbc={}] start_peering_interval up [3,1,5] -> [1,2,3], acting [3,1,5] -> [1,2,3], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.15( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877305984s) [1,2,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283691406s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.14( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.871586800s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278198242s@ mbc={}] start_peering_interval up [1,3,2] -> [3,4,2], acting [1,3,2] -> [3,4,2], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.14( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.871586800s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.278198242s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.16( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877406120s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.284179688s@ mbc={}] start_peering_interval up [3,1,5] -> [0,1,2], acting [3,1,5] -> [0,1,2], acting_primary 3 -> 0, up_primary 3 -> 0, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.14( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877151489s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283935547s@ mbc={}] start_peering_interval up [3,1,5] -> [3,4,2], acting [3,1,5] -> [3,4,2], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.14( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877151489s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.283935547s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.16( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.877290726s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.284179688s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.15( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.871500015s) [1,3,5] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278320312s@ mbc={}] start_peering_interval up [1,3,2] -> [1,3,5], acting [1,3,2] -> [1,3,5], acting_primary 1 -> 1, up_primary 1 -> 1, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.15( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.871452332s) [1,3,5] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278320312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.14( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887166977s) [2,4,0] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294189453s@ mbc={}] start_peering_interval up [4,5,3] -> [2,4,0], acting [4,5,3] -> [2,4,0], acting_primary 4 -> 2, up_primary 4 -> 2, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.14( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887148857s) [2,4,0] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.294189453s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.12( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870994568s) [2,3,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278198242s@ mbc={}] start_peering_interval up [1,3,2] -> [2,3,1], acting [1,3,2] -> [2,3,1], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.12( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870978355s) [2,3,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278198242s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.13( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.876385689s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283691406s@ mbc={}] start_peering_interval up [3,1,5] -> [1,0,2], acting [3,1,5] -> [1,0,2], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.15( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882670403s) [3,2,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.290039062s@ mbc={}] start_peering_interval up [4,5,3] -> [3,2,1], acting [4,5,3] -> [3,2,1], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.13( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870937347s) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278320312s@ mbc={}] start_peering_interval up [1,3,2] -> [3,2,1], acting [1,3,2] -> [3,2,1], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.13( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870937347s) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.278320312s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.13( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.876276970s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283691406s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.15( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882670403s) [3,2,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.290039062s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.16( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886488914s) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294067383s@ mbc={}] start_peering_interval up [4,5,3] -> [3,5,1], acting [4,5,3] -> [3,5,1], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.16( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886488914s) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.294067383s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.12( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.876004219s) [4,2,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283569336s@ mbc={}] start_peering_interval up [3,1,5] -> [4,2,3], acting [3,1,5] -> [4,2,3], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.12( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.875979424s) [4,2,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283569336s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.11( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870247841s) [0,2,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277954102s@ mbc={}] start_peering_interval up [3,1,5] -> [0,2,4], acting [3,1,5] -> [0,2,4], acting_primary 3 -> 0, up_primary 3 -> 0, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.11( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870226860s) [0,2,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277954102s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.17( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886143684s) [5,1,0] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293945312s@ mbc={}] start_peering_interval up [4,5,3] -> [5,1,0], acting [4,5,3] -> [5,1,0], acting_primary 4 -> 5, up_primary 4 -> 5, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.17( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886126518s) [5,1,0] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293945312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.11( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870516777s) [4,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278442383s@ mbc={}] start_peering_interval up [1,3,2] -> [4,0,2], acting [1,3,2] -> [4,0,2], acting_primary 1 -> 4, up_primary 1 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.11( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.870463371s) [4,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278442383s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.10( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869736671s) [3,1,5] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [1,3,2] -> [3,1,5], acting [1,3,2] -> [3,1,5], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.8( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886875153s) [1,2,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295043945s@ mbc={}] start_peering_interval up [4,5,3] -> [1,2,3], acting [4,5,3] -> [1,2,3], acting_primary 4 -> 1, up_primary 4 -> 1, role 2 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869709969s) [4,0,5] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [3,1,5] -> [4,0,5], acting [3,1,5] -> [4,0,5], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869692802s) [4,0,5] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277832031s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.8( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886816978s) [1,2,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295043945s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.10( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869967461s) [4,5,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278198242s@ mbc={}] start_peering_interval up [3,1,5] -> [4,5,3], acting [3,1,5] -> [4,5,3], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.9( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887157440s) [0,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295410156s@ mbc={}] start_peering_interval up [4,5,3] -> [0,4,2], acting [4,5,3] -> [0,4,2], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.9( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.887139320s) [0,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295410156s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.10( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869906425s) [4,5,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278198242s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.10( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869736671s) [3,1,5] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277832031s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.f( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869658470s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277954102s@ mbc={}] start_peering_interval up [1,3,2] -> [0,1,2], acting [1,3,2] -> [0,1,2], acting_primary 1 -> 0, up_primary 1 -> 0, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.f( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869642258s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277954102s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.a( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886777878s) [2,1,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295288086s@ mbc={}] start_peering_interval up [4,5,3] -> [2,1,3], acting [4,5,3] -> [2,1,3], acting_primary 4 -> 2, up_primary 4 -> 2, role 2 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.a( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.886763573s) [2,1,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295288086s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.c( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869159698s) [0,5,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [1,3,2] -> [0,5,1], acting [1,3,2] -> [0,5,1], acting_primary 1 -> 0, up_primary 1 -> 0, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.c( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869145393s) [0,5,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277832031s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869296074s) [1,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278076172s@ mbc={}] start_peering_interval up [3,1,5] -> [1,5,0], acting [3,1,5] -> [1,5,0], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869277954s) [1,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278076172s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868715286s) [1,0,5] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277587891s@ mbc={}] start_peering_interval up [3,1,5] -> [1,0,5], acting [3,1,5] -> [1,0,5], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.c( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.885135651s) [5,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293945312s@ mbc={}] start_peering_interval up [4,5,3] -> [5,0,4], acting [4,5,3] -> [5,0,4], acting_primary 4 -> 5, up_primary 4 -> 5, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.c( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.885119438s) [5,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293945312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.d( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868892670s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [1,3,2] -> [3,1,2], acting [1,3,2] -> [3,1,2], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868654251s) [1,0,5] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277587891s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868336678s) [1,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277343750s@ mbc={}] start_peering_interval up [3,1,5] -> [1,3,2], acting [3,1,5] -> [1,3,2], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868314743s) [1,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277343750s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.d( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868892670s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277832031s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.a( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868137360s) [3,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277343750s@ mbc={}] start_peering_interval up [1,3,2] -> [3,5,1], acting [1,3,2] -> [3,5,1], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.a( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868137360s) [3,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277343750s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867900848s) [1,5,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277221680s@ mbc={}] start_peering_interval up [3,1,5] -> [1,5,3], acting [3,1,5] -> [1,5,3], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.d( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.885049820s) [4,0,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294433594s@ mbc={}] start_peering_interval up [4,5,3] -> [4,0,2], acting [4,5,3] -> [4,0,2], acting_primary 4 -> 4, up_primary 4 -> 4, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867877007s) [1,5,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277221680s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.d( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.884968758s) [4,0,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.294433594s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.2( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867770195s) [5,4,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277709961s@ mbc={}] start_peering_interval up [1,3,2] -> [5,4,0], acting [1,3,2] -> [5,4,0], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.2( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867751122s) [5,4,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277709961s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.3( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867812157s) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277709961s@ mbc={}] start_peering_interval up [3,1,5] -> [3,2,1], acting [3,1,5] -> [3,2,1], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.3( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867812157s) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277709961s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.7( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.885532379s) [3,5,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295532227s@ mbc={}] start_peering_interval up [4,5,3] -> [3,5,4], acting [4,5,3] -> [3,5,4], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.b( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867350578s) [2,0,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277099609s@ mbc={}] start_peering_interval up [1,3,2] -> [2,0,1], acting [1,3,2] -> [2,0,1], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.b( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866900444s) [2,0,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277099609s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.4( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.884249687s) [1,2,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293945312s@ mbc={}] start_peering_interval up [4,5,3] -> [1,2,3], acting [4,5,3] -> [1,2,3], acting_primary 4 -> 1, up_primary 4 -> 1, role 2 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867418289s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277587891s@ mbc={}] start_peering_interval up [1,3,2] -> [2,3,4], acting [1,3,2] -> [2,3,4], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867379189s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277587891s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.7( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.885532379s) [3,5,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.295532227s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.4( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.883658409s) [1,2,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293945312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.7( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866987228s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [3,1,5] -> [3,1,2], acting [3,1,5] -> [3,1,2], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868487358s) [2,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278198242s@ mbc={}] start_peering_interval up [3,1,5] -> [2,4,3], acting [3,1,5] -> [2,4,3], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867117882s) [2,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277709961s@ mbc={}] start_peering_interval up [3,1,5] -> [2,4,3], acting [3,1,5] -> [2,4,3], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.6( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.884590149s) [2,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295166016s@ mbc={}] start_peering_interval up [4,5,3] -> [2,0,4], acting [4,5,3] -> [2,0,4], acting_primary 4 -> 2, up_primary 4 -> 2, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.7( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866987228s) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277465820s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867645264s) [2,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278198242s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.6( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.884507179s) [2,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295166016s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867006302s) [2,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277709961s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.6( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.867042542s) [5,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277954102s@ mbc={}] start_peering_interval up [1,3,2] -> [5,0,4], acting [1,3,2] -> [5,0,4], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.6( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866943359s) [5,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277954102s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.5( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.883123398s) [5,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.294433594s@ mbc={}] start_peering_interval up [4,5,3] -> [5,3,1], acting [4,5,3] -> [5,3,1], acting_primary 4 -> 5, up_primary 4 -> 5, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.2( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866553307s) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277832031s@ mbc={}] start_peering_interval up [3,1,5] -> [3,5,4], acting [3,1,5] -> [3,5,4], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.5( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.883066177s) [5,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.294433594s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.3( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866168022s) [0,2,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [1,3,2] -> [0,2,4], acting [1,3,2] -> [0,2,4], acting_primary 1 -> 0, up_primary 1 -> 0, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.3( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866049767s) [0,2,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277465820s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.5( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865891457s) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [1,3,2] -> [3,5,4], acting [1,3,2] -> [3,5,4], acting_primary 1 -> 3, up_primary 1 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.4( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865903854s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [3,1,5] -> [5,1,3], acting [3,1,5] -> [5,1,3], acting_primary 3 -> 5, up_primary 3 -> 5, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.5( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865891457s) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277465820s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.2( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.866553307s) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.277832031s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.4( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865838051s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277465820s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.2( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880931854s) [0,5,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.292846680s@ mbc={}] start_peering_interval up [4,5,3] -> [0,5,1], acting [4,5,3] -> [0,5,1], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.2( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880848885s) [0,5,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.292846680s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.5( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865355492s) [1,2,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [3,1,5] -> [1,2,0], acting [3,1,5] -> [1,2,0], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.4( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865439415s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277587891s@ mbc={}] start_peering_interval up [1,3,2] -> [2,1,3], acting [1,3,2] -> [2,1,3], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.5( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865276337s) [1,2,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277465820s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.4( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865383148s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277587891s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.3( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.883850098s) [2,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295288086s@ mbc={}] start_peering_interval up [4,5,3] -> [2,3,1], acting [4,5,3] -> [2,3,1], acting_primary 4 -> 2, up_primary 4 -> 2, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.6( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864792824s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277343750s@ mbc={}] start_peering_interval up [3,1,5] -> [2,1,3], acting [3,1,5] -> [2,1,3], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.f( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881113052s) [0,1,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293701172s@ mbc={}] start_peering_interval up [4,5,3] -> [0,1,2], acting [4,5,3] -> [0,1,2], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.3( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882793427s) [2,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295288086s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.6( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864727020s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277343750s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.f( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881065369s) [0,1,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293701172s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.7( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864756584s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277465820s@ mbc={}] start_peering_interval up [1,3,2] -> [2,1,3], acting [1,3,2] -> [2,1,3], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.7( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864673615s) [2,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277465820s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.8( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865064621s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277954102s@ mbc={}] start_peering_interval up [3,1,5] -> [0,1,2], acting [3,1,5] -> [0,1,2], acting_primary 3 -> 0, up_primary 3 -> 0, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.8( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.865013123s) [0,1,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277954102s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.9( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864226341s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277343750s@ mbc={}] start_peering_interval up [1,3,2] -> [4,3,2], acting [1,3,2] -> [4,3,2], acting_primary 1 -> 4, up_primary 1 -> 4, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.9( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864097595s) [5,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277343750s@ mbc={}] start_peering_interval up [3,1,5] -> [5,0,4], acting [3,1,5] -> [5,0,4], acting_primary 3 -> 5, up_primary 3 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882295609s) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295532227s@ mbc={}] start_peering_interval up [4,5,3] -> [3,5,1], acting [4,5,3] -> [3,5,1], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.9( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864117622s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277343750s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.8( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863845825s) [4,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277221680s@ mbc={}] start_peering_interval up [1,3,2] -> [4,5,0], acting [1,3,2] -> [4,5,0], acting_primary 1 -> 4, up_primary 1 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.9( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.864036560s) [5,0,4] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277343750s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.8( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863790512s) [4,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277221680s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1d( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882502556s) [5,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295898438s@ mbc={}] start_peering_interval up [4,5,3] -> [5,0,4], acting [4,5,3] -> [5,0,4], acting_primary 4 -> 5, up_primary 4 -> 5, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1d( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882462502s) [5,0,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295898438s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869787216s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283447266s@ mbc={}] start_peering_interval up [3,1,5] -> [3,4,2], acting [3,1,5] -> [3,4,2], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1b( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863435745s) [4,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277099609s@ mbc={}] start_peering_interval up [1,3,2] -> [4,0,2], acting [1,3,2] -> [4,0,2], acting_primary 1 -> 4, up_primary 1 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1a( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.869787216s) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1187.283447266s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1c( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881690979s) [0,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295532227s@ mbc={}] start_peering_interval up [4,5,3] -> [0,4,2], acting [4,5,3] -> [0,4,2], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1c( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881639481s) [0,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295532227s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863991737s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278198242s@ mbc={}] start_peering_interval up [3,1,5] -> [1,0,2], acting [3,1,5] -> [1,0,2], acting_primary 3 -> 1, up_primary 3 -> 1, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882295609s) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1189.295532227s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1b( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.862690926s) [4,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277099609s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1b( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863902092s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278198242s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1b( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879373550s) [0,5,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293945312s@ mbc={}] start_peering_interval up [4,5,3] -> [0,5,4], acting [4,5,3] -> [0,5,4], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1b( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879321098s) [0,5,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293945312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1d( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863752365s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.278320312s@ mbc={}] start_peering_interval up [1,3,2] -> [1,0,2], acting [1,3,2] -> [1,0,2], acting_primary 1 -> 1, up_primary 1 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1a( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880701065s) [1,0,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295532227s@ mbc={}] start_peering_interval up [4,5,3] -> [1,0,5], acting [4,5,3] -> [1,0,5], acting_primary 4 -> 1, up_primary 4 -> 1, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868832588s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283569336s@ mbc={}] start_peering_interval up [3,1,5] -> [4,3,2], acting [3,1,5] -> [4,3,2], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.1a( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880642891s) [1,0,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295532227s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1c( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868772507s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283569336s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1c( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.861900330s) [0,5,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.276733398s@ mbc={}] start_peering_interval up [1,3,2] -> [0,5,1], acting [1,3,2] -> [0,5,1], acting_primary 1 -> 0, up_primary 1 -> 0, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868812561s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283813477s@ mbc={}] start_peering_interval up [3,1,5] -> [4,3,2], acting [3,1,5] -> [4,3,2], acting_primary 3 -> 4, up_primary 3 -> 4, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1d( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.863420486s) [1,0,2] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.278320312s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1d( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868758202s) [4,3,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283813477s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1c( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.861574173s) [0,5,1] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.276733398s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1f( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.861671448s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.277099609s@ mbc={}] start_peering_interval up [1,3,2] -> [2,3,4], acting [1,3,2] -> [2,3,4], acting_primary 1 -> 2, up_primary 1 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.19( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878316879s) [2,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.293823242s@ mbc={}] start_peering_interval up [4,5,3] -> [2,3,1], acting [4,5,3] -> [2,3,1], acting_primary 4 -> 2, up_primary 4 -> 2, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1f( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.861620903s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.277099609s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.18( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880105019s) [0,1,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1189.295654297s@ mbc={}] start_peering_interval up [4,5,3] -> [0,1,5], acting [4,5,3] -> [0,1,5], acting_primary 4 -> 0, up_primary 4 -> 0, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868332863s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283813477s@ mbc={}] start_peering_interval up [3,1,5] -> [5,1,3], acting [3,1,5] -> [5,1,3], acting_primary 3 -> 5, up_primary 3 -> 5, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.19( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878263474s) [2,3,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.293823242s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1e( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868195534s) [5,1,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283813477s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868112564s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.283935547s@ mbc={}] start_peering_interval up [3,1,5] -> [2,3,4], acting [3,1,5] -> [2,3,4], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[2.1f( empty local-lis/les=36/37 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.868058205s) [2,3,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.283935547s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1e( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.860959053s) [5,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.276855469s@ mbc={}] start_peering_interval up [1,3,2] -> [5,4,3], acting [1,3,2] -> [5,4,3], acting_primary 1 -> 5, up_primary 1 -> 5, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[5.18( empty local-lis/les=38/39 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880055428s) [0,1,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1189.295654297s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1e( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.860768318s) [5,4,3] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.276855469s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.9( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.f( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.3( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.2( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.c( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.f( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,1,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.8( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1a( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.860086441s) [4,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active pruub 1187.276733398s@ mbc={}] start_peering_interval up [1,3,2] -> [4,5,0], acting [1,3,2] -> [4,5,0], acting_primary 1 -> 4, up_primary 1 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[3.1a( empty local-lis/les=36/37 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42 pruub=8.859960556s) [4,5,0] r=-1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1187.276733398s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.16( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.10( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.11( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.17( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.883067131s) [5,3,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758666992s@ mbc={}] start_peering_interval up [5,0,1] -> [5,3,4], acting [5,0,1] -> [5,3,4], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1a( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.946166039s) [3,4,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822143555s@ mbc={}] start_peering_interval up [5,0,1] -> [3,4,5], acting [5,0,1] -> [3,4,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.14( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940663338s) [5,3,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816772461s@ mbc={}] start_peering_interval up [5,0,1] -> [5,3,4], acting [5,0,1] -> [5,3,4], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.15( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940402031s) [3,4,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816528320s@ mbc={}] start_peering_interval up [5,0,1] -> [3,4,5], acting [5,0,1] -> [3,4,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.18( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881170273s) [3,2,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.757324219s@ mbc={}] start_peering_interval up [5,0,1] -> [3,2,1], acting [5,0,1] -> [3,2,1], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.18( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881101608s) [3,2,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.757324219s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1a( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.945845604s) [3,4,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822143555s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.17( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.882282257s) [5,3,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758666992s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.15( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881757736s) [4,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758300781s@ mbc={}] start_peering_interval up [5,0,1] -> [4,2,0], acting [5,0,1] -> [4,2,0], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.15( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881709099s) [4,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758300781s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.15( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940068245s) [3,4,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816528320s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.18( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,2,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.14( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880020142s) [4,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.756835938s@ mbc={}] start_peering_interval up [5,0,1] -> [4,2,0], acting [5,0,1] -> [4,2,0], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.14( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879970551s) [4,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.756835938s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.16( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939699173s) [5,0,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816650391s@ mbc={}] start_peering_interval up [5,0,1] -> [5,0,4], acting [5,0,1] -> [5,0,4], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.16( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939658165s) [5,0,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816650391s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.13( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881719589s) [0,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758911133s@ mbc={}] start_peering_interval up [5,0,1] -> [0,1,5], acting [5,0,1] -> [0,1,5], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.14( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939601898s) [5,3,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816772461s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.13( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.881719589s) [0,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1193.758911133s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.17( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939104080s) [1,2,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816406250s@ mbc={}] start_peering_interval up [5,0,1] -> [1,2,0], acting [5,0,1] -> [1,2,0], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.17( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939062119s) [1,2,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816406250s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.11( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939034462s) [5,1,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816528320s@ mbc={}] start_peering_interval up [5,0,1] -> [5,1,3], acting [5,0,1] -> [5,1,3], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.11( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939005852s) [5,1,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816528320s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.10( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939322472s) [2,1,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816894531s@ mbc={}] start_peering_interval up [5,0,1] -> [2,1,3], acting [5,0,1] -> [2,1,3], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.11( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879367828s) [2,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.756958008s@ mbc={}] start_peering_interval up [5,0,1] -> [2,0,4], acting [5,0,1] -> [2,0,4], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.10( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939271927s) [2,1,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816894531s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.12( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879326820s) [5,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.756958008s@ mbc={}] start_peering_interval up [5,0,1] -> [5,1,3], acting [5,0,1] -> [5,1,3], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.11( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879333496s) [2,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.756958008s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.12( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879269600s) [5,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.756958008s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.12( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938502312s) [4,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816406250s@ mbc={}] start_peering_interval up [5,0,1] -> [4,0,2], acting [5,0,1] -> [4,0,2], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.10( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880512238s) [5,4,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758422852s@ mbc={}] start_peering_interval up [5,0,1] -> [5,4,3], acting [5,0,1] -> [5,4,3], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.12( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938468933s) [4,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816406250s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.10( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880465508s) [5,4,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758422852s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.f( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880390167s) [2,4,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758422852s@ mbc={}] start_peering_interval up [5,0,1] -> [2,4,3], acting [5,0,1] -> [2,4,3], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.f( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880357742s) [2,4,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758422852s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.d( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938950539s) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.817138672s@ mbc={}] start_peering_interval up [5,0,1] -> [0,5,1], acting [5,0,1] -> [0,5,1], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.d( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938950539s) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1195.817138672s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.1a( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,4,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.c( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.943573952s) [2,0,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822021484s@ mbc={}] start_peering_interval up [5,0,1] -> [2,0,4], acting [5,0,1] -> [2,0,4], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.e( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880349159s) [3,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758666992s@ mbc={}] start_peering_interval up [5,0,1] -> [3,4,2], acting [5,0,1] -> [3,4,2], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.e( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880127907s) [3,4,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758666992s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.d( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880194664s) [3,1,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758789062s@ mbc={}] start_peering_interval up [5,0,1] -> [3,1,5], acting [5,0,1] -> [3,1,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.d( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.880141258s) [3,1,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758789062s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.13( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938136101s) [2,4,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816772461s@ mbc={}] start_peering_interval up [5,0,1] -> [2,4,3], acting [5,0,1] -> [2,4,3], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.f( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.942028046s) [5,4,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.820678711s@ mbc={}] start_peering_interval up [5,0,1] -> [5,4,0], acting [5,0,1] -> [5,4,0], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.f( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941985130s) [5,4,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.820678711s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.c( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879609108s) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758544922s@ mbc={}] start_peering_interval up [5,0,1] -> [0,5,1], acting [5,0,1] -> [0,5,1], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.e( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937836647s) [3,2,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.816772461s@ mbc={}] start_peering_interval up [5,0,1] -> [3,2,4], acting [5,0,1] -> [3,2,4], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.c( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879609108s) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1193.758544922s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.e( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937806129s) [3,2,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816772461s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.b( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879615784s) [2,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758789062s@ mbc={}] start_peering_interval up [5,0,1] -> [2,0,4], acting [5,0,1] -> [2,0,4], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.b( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879582405s) [2,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758789062s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.c( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.943537712s) [2,0,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822021484s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.9( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937443733s) [5,1,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.817016602s@ mbc={}] start_peering_interval up [5,0,1] -> [5,1,0], acting [5,0,1] -> [5,1,0], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.9( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937380791s) [5,1,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.817016602s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.2( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878757477s) [1,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758789062s@ mbc={}] start_peering_interval up [5,0,1] -> [1,3,2], acting [5,0,1] -> [1,3,2], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.2( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878665924s) [1,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758789062s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.2( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941784859s) [3,5,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.821655273s@ mbc={}] start_peering_interval up [5,0,1] -> [3,5,4], acting [5,0,1] -> [3,5,4], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.13( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937134743s) [2,4,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.816772461s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.2( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941484451s) [3,5,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.821655273s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.3( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879460335s) [1,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759765625s@ mbc={}] start_peering_interval up [5,0,1] -> [1,3,2], acting [5,0,1] -> [1,3,2], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.3( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.879404068s) [1,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759765625s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.942155838s) [1,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822509766s@ mbc={}] start_peering_interval up [5,0,1] -> [1,0,2], acting [5,0,1] -> [1,0,2], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.942104340s) [1,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822509766s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878440857s) [4,0,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.758911133s@ mbc={}] start_peering_interval up [5,0,1] -> [4,0,2], acting [5,0,1] -> [4,0,2], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.3( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941740036s) [0,4,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822265625s@ mbc={}] start_peering_interval up [5,0,1] -> [0,4,2], acting [5,0,1] -> [0,4,2], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.3( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941740036s) [0,4,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1195.822265625s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1b( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941103935s) [1,3,2] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.821777344s@ mbc={}] start_peering_interval up [5,0,1] -> [1,3,2], acting [5,0,1] -> [1,3,2], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.19( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878316879s) [1,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759033203s@ mbc={}] start_peering_interval up [5,0,1] -> [1,2,0], acting [5,0,1] -> [1,2,0], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1b( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941056252s) [1,3,2] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.821777344s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.19( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878287315s) [1,2,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759033203s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.6( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941495895s) [2,3,1] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822509766s@ mbc={}] start_peering_interval up [5,0,1] -> [2,3,1], acting [5,0,1] -> [2,3,1], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.9( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878192902s) [1,5,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759399414s@ mbc={}] start_peering_interval up [5,0,1] -> [1,5,0], acting [5,0,1] -> [1,5,0], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.9( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878054619s) [1,5,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759399414s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.6( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.941408157s) [2,3,1] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822509766s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.b( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940237999s) [2,3,1] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.821655273s@ mbc={}] start_peering_interval up [5,0,1] -> [2,3,1], acting [5,0,1] -> [2,3,1], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1a( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.877655983s) [3,2,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759155273s@ mbc={}] start_peering_interval up [5,0,1] -> [3,2,4], acting [5,0,1] -> [3,2,4], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.b( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940200806s) [2,3,1] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.821655273s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1a( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.877587318s) [3,2,4] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759155273s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.18( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940244675s) [5,3,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822021484s@ mbc={}] start_peering_interval up [5,0,1] -> [5,3,4], acting [5,0,1] -> [5,3,4], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.18( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.940200806s) [5,3,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822021484s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.5( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.877901077s) [3,1,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759765625s@ mbc={}] start_peering_interval up [5,0,1] -> [3,1,2], acting [5,0,1] -> [3,1,2], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.5( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.877847672s) [3,1,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759765625s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.15( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,4,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.7( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939905167s) [3,5,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822143555s@ mbc={}] start_peering_interval up [5,0,1] -> [3,5,4], acting [5,0,1] -> [3,5,4], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.7( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.939857483s) [3,5,4] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822143555s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.e( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.a( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876685143s) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759399414s@ mbc={}] start_peering_interval up [5,0,1] -> [0,5,1], acting [5,0,1] -> [0,5,1], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.a( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876685143s) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1193.759399414s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.8( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.934108734s) [3,1,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.817016602s@ mbc={}] start_peering_interval up [5,0,1] -> [3,1,5], acting [5,0,1] -> [3,1,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.8( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.934064865s) [3,1,5] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.817016602s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.878399849s) [4,0,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.758911133s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.d( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.6( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876388550s) [4,5,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759643555s@ mbc={}] start_peering_interval up [5,0,1] -> [4,5,0], acting [5,0,1] -> [4,5,0], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.6( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876343727s) [4,5,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759643555s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.7( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875586510s) [2,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759033203s@ mbc={}] start_peering_interval up [5,0,1] -> [2,1,3], acting [5,0,1] -> [2,1,3], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.7( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875512123s) [2,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759033203s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.5( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938735008s) [0,1,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822387695s@ mbc={}] start_peering_interval up [5,0,1] -> [0,1,2], acting [5,0,1] -> [0,1,2], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.5( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938735008s) [0,1,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1195.822387695s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.8( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875763893s) [1,3,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759521484s@ mbc={}] start_peering_interval up [5,0,1] -> [1,3,5], acting [5,0,1] -> [1,3,5], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.8( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875670433s) [1,3,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759521484s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.a( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938765526s) [0,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822753906s@ mbc={}] start_peering_interval up [5,0,1] -> [0,5,4], acting [5,0,1] -> [0,5,4], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1b( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876098633s) [3,5,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.760131836s@ mbc={}] start_peering_interval up [5,0,1] -> [3,5,1], acting [5,0,1] -> [3,5,1], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.a( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.938765526s) [0,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1195.822753906s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1b( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.876063347s) [3,5,1] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.760131836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.e( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,2,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.4( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875746727s) [5,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.760009766s@ mbc={}] start_peering_interval up [5,0,1] -> [5,1,3], acting [5,0,1] -> [5,1,3], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.19( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937700272s) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822021484s@ mbc={}] start_peering_interval up [5,0,1] -> [0,5,1], acting [5,0,1] -> [0,5,1], acting_primary 5 -> 0, up_primary 5 -> 0, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.19( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937700272s) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown pruub 1195.822021484s@ mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1e( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937514305s) [4,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822021484s@ mbc={}] start_peering_interval up [5,0,1] -> [4,0,2], acting [5,0,1] -> [4,0,2], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1c( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875411034s) [1,2,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.759887695s@ mbc={}] start_peering_interval up [5,0,1] -> [1,2,3], acting [5,0,1] -> [1,2,3], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.4( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875664711s) [5,1,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.760009766s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1e( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937453270s) [4,0,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822021484s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1f( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937765121s) [5,4,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822387695s@ mbc={}] start_peering_interval up [5,0,1] -> [5,4,0], acting [5,0,1] -> [5,4,0], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1f( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937730789s) [5,4,0] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822387695s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1d( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875315666s) [4,3,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.760131836s@ mbc={}] start_peering_interval up [5,0,1] -> [4,3,5], acting [5,0,1] -> [4,3,5], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1c( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875371933s) [1,2,3] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.759887695s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1d( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.875247002s) [4,3,5] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.760131836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1d( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937389374s) [5,4,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822387695s@ mbc={}] start_peering_interval up [5,0,1] -> [5,4,3], acting [5,0,1] -> [5,4,3], acting_primary 5 -> 5, up_primary 5 -> 5, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.2( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1d( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937347412s) [5,4,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822387695s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1e( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.874891281s) [2,0,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.760131836s@ mbc={}] start_peering_interval up [5,0,1] -> [2,0,1], acting [5,0,1] -> [2,0,1], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1e( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.874861717s) [2,0,1] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.760131836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.1c( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1c( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937325478s) [1,5,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active pruub 1195.822631836s@ mbc={}] start_peering_interval up [5,0,1] -> [1,5,3], acting [5,0,1] -> [1,5,3], acting_primary 5 -> 1, up_primary 5 -> 1, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[6.1c( empty local-lis/les=40/41 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42 pruub=12.937263489s) [1,5,3] r=-1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822631836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.1b( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.1c( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.1a( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,2,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1f( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.874391556s) [4,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active pruub 1193.760253906s@ mbc={}] start_peering_interval up [5,0,1] -> [4,3,2], acting [5,0,1] -> [4,3,2], acting_primary 5 -> 4, up_primary 5 -> 4, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.18( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[4.1f( empty local-lis/les=38/39 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42 pruub=10.873375893s) [4,3,2] r=-1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown NOTIFY pruub 1193.760253906s@ mbc={}] state: transitioning to Stray Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.5( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,1,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.7( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.8( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,1,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:34 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.1b( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.13( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,0,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.11( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [5,1,3] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.19( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [5,0,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.12( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,1,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.d( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,0,5] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.c( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,5,0] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.10( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,4,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.1a( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [4,5,0] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.5( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,2,0] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.17( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,3,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.1d( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,0,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.14( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [5,3,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.1b( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [4,0,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.1b( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [1,0,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.f( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [4,0,5] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.1d( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [4,3,5] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.8( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [4,5,0] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.1f( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [4,3,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.1a( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [1,0,5] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.d( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [4,0,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.1d( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [5,4,3] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.11( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [4,0,2] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.4( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,1,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.18( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [5,3,4] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.b( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [2,0,1] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.6( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [2,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.1b( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [1,3,2] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.14( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [2,4,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.19( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [2,0,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.1c( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [1,2,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.2( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [1,3,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.17( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,1,0] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.c( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.3( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [1,3,2] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.8( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [1,3,5] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[2.9( empty local-lis/les=0/0 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [5,0,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.6( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [5,0,4] r=1 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.1c( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [1,5,3] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[3.2( empty local-lis/les=0/0 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [5,4,0] r=2 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.11( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,4,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 42 pg[5.1d( empty local-lis/les=0/0 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [5,0,4] r=1 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.2( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.15( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,4,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.7( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.2( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.1a( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,4,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.10( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[6.a( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [0,5,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.1b( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.7( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,5,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.5( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,5,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.18( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[3.c( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.16( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[4.c( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[4.13( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.a( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.8( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,1,5] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.10( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,1,5] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.1( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.1b( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.d( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,1,5] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[4.a( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.2( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,5,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.13( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [2,4,3] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[6.19( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[3.1c( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,5,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[6.d( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [0,5,1] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.10( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [2,1,3] r=2 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.b( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [2,3,1] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.7( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [2,1,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[6.6( empty local-lis/les=0/0 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [2,3,1] r=1 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[3.f( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.9( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.1c( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 42 pg[4.f( empty local-lis/les=0/0 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [2,4,3] r=2 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[6.e( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [3,2,4] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.1a( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,2,4] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.e( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.7( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.3( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.d( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.5( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,1,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.14( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.13( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,2,1] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[4.18( empty local-lis/les=42/43 n=0 ec=38/21 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,2,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.14( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[3.16( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.17( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.1f( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,4,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[2.1a( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [3,4,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[32722]: osd.3 pg_epoch: 43 pg[5.15( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [3,2,1] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[5.f( empty local-lis/les=42/43 n=0 ec=38/23 lis/c=38/38 les/c/f=39/39/0 sis=42) [0,1,2] r=0 lpr=42 pi=[38,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[2.8( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[3.3( empty local-lis/les=42/43 n=0 ec=36/19 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[2.11( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,2,4] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[2.16( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=36/36 les/c/f=37/37/0 sis=42) [0,1,2] r=0 lpr=42 pi=[36,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[6.5( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [0,1,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost ceph-osd[31760]: osd.0 pg_epoch: 43 pg[6.3( empty local-lis/les=42/43 n=0 ec=40/29 lis/c=40/40 les/c/f=41/41/0 sis=42) [0,4,2] r=0 lpr=42 pi=[40,42)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:07:35 localhost python3[57590]: ansible-file Invoked with path=/var/lib/tripleo-config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.6( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.644747734s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1195.822143555s@ mbc={}] start_peering_interval up [0,1,5] -> [2,1,0], acting [0,1,5] -> [2,1,0], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.a( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.644117355s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 active+scrubbing pruub 1195.821533203s@ [ 7.a: ] mbc={}] start_peering_interval up [0,1,5] -> [2,1,0], acting [0,1,5] -> [2,1,0], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.2( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.644856453s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1195.822265625s@ mbc={}] start_peering_interval up [0,1,5] -> [2,1,0], acting [0,1,5] -> [2,1,0], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.6( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.644654274s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822143555s@ mbc={}] state: transitioning to Stray Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.2( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.644773483s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1195.822265625s@ mbc={}] state: transitioning to Stray Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.643900871s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1195.821777344s@ mbc={}] start_peering_interval up [0,1,5] -> [2,1,0], acting [0,1,5] -> [2,1,0], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.643785477s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1195.821777344s@ mbc={}] state: transitioning to Stray Nov 27 03:07:36 localhost ceph-osd[31760]: osd.0 pg_epoch: 44 pg[7.a( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=44 pruub=10.643353462s) [2,1,0] r=2 lpr=44 pi=[40,44)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1195.821533203s@ mbc={}] state: transitioning to Stray Nov 27 03:07:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:07:36 localhost systemd[1]: tmp-crun.6921i0.mount: Deactivated successfully. Nov 27 03:07:36 localhost podman[57591]: 2025-11-27 08:07:36.976549127 +0000 UTC m=+0.074070701 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, architecture=x86_64, io.openshift.expose-services=, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, url=https://www.redhat.com, managed_by=tripleo_ansible, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team) Nov 27 03:07:37 localhost podman[57591]: 2025-11-27 08:07:37.194337354 +0000 UTC m=+0.291858978 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, release=1761123044, vcs-type=git, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:07:37 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:07:37 localhost python3[57634]: ansible-file Invoked with path=/var/lib/tripleo-config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:39 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.0 deep-scrub starts Nov 27 03:07:40 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.13 scrub starts Nov 27 03:07:41 localhost python3[57682]: ansible-ansible.legacy.stat Invoked with path=/var/lib/tripleo-config/ceph/ceph.client.openstack.keyring follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:07:41 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.d scrub starts Nov 27 03:07:41 localhost python3[57725]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230861.0461998-91943-157464963771786/source dest=/var/lib/tripleo-config/ceph/ceph.client.openstack.keyring mode=600 _original_basename=ceph.client.openstack.keyring follow=False checksum=4387bc148c9a58f9a0b1eb236093f27263bcb30f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:42 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.17 deep-scrub starts Nov 27 03:07:43 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.8 scrub starts Nov 27 03:07:43 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.8 scrub ok Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.244102478s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.821777344s@ mbc={}] start_peering_interval up [0,1,5] -> [5,3,4], acting [0,1,5] -> [5,3,4], acting_primary 0 -> 5, up_primary 0 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.b( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.245376587s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.822998047s@ mbc={}] start_peering_interval up [0,1,5] -> [5,3,4], acting [0,1,5] -> [5,3,4], acting_primary 0 -> 5, up_primary 0 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.244007111s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.821777344s@ mbc={}] state: transitioning to Stray Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.b( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.245240211s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.822998047s@ mbc={}] state: transitioning to Stray Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.3( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.244223595s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.822021484s@ mbc={}] start_peering_interval up [0,1,5] -> [5,3,4], acting [0,1,5] -> [5,3,4], acting_primary 0 -> 5, up_primary 0 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.243796349s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.821899414s@ mbc={}] start_peering_interval up [0,1,5] -> [5,3,4], acting [0,1,5] -> [5,3,4], acting_primary 0 -> 5, up_primary 0 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.3( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.243852615s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.822021484s@ mbc={}] state: transitioning to Stray Nov 27 03:07:44 localhost ceph-osd[31760]: osd.0 pg_epoch: 46 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46 pruub=11.243581772s) [5,3,4] r=-1 lpr=46 pi=[40,46)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.821899414s@ mbc={}] state: transitioning to Stray Nov 27 03:07:44 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.c scrub starts Nov 27 03:07:44 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.c scrub ok Nov 27 03:07:45 localhost ceph-osd[32722]: osd.3 pg_epoch: 46 pg[7.7( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46) [5,3,4] r=1 lpr=46 pi=[40,46)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:45 localhost ceph-osd[32722]: osd.3 pg_epoch: 46 pg[7.3( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46) [5,3,4] r=1 lpr=46 pi=[40,46)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:45 localhost ceph-osd[32722]: osd.3 pg_epoch: 46 pg[7.f( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46) [5,3,4] r=1 lpr=46 pi=[40,46)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:45 localhost ceph-osd[32722]: osd.3 pg_epoch: 46 pg[7.b( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=46) [5,3,4] r=1 lpr=46 pi=[40,46)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:46 localhost ceph-osd[31760]: osd.0 pg_epoch: 48 pg[7.4( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48 pruub=8.652174950s) [2,3,4] r=-1 lpr=48 pi=[40,48)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.822631836s@ mbc={}] start_peering_interval up [0,1,5] -> [2,3,4], acting [0,1,5] -> [2,3,4], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:46 localhost ceph-osd[31760]: osd.0 pg_epoch: 48 pg[7.4( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48 pruub=8.652100563s) [2,3,4] r=-1 lpr=48 pi=[40,48)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.822631836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:46 localhost ceph-osd[31760]: osd.0 pg_epoch: 48 pg[7.c( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48 pruub=8.651227951s) [2,3,4] r=-1 lpr=48 pi=[40,48)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1203.821899414s@ mbc={}] start_peering_interval up [0,1,5] -> [2,3,4], acting [0,1,5] -> [2,3,4], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:46 localhost ceph-osd[31760]: osd.0 pg_epoch: 48 pg[7.c( v 33'39 (0'0,33'39] local-lis/les=40/41 n=1 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48 pruub=8.651105881s) [2,3,4] r=-1 lpr=48 pi=[40,48)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1203.821899414s@ mbc={}] state: transitioning to Stray Nov 27 03:07:46 localhost python3[57787]: ansible-ansible.legacy.stat Invoked with path=/var/lib/tripleo-config/ceph/ceph.client.manila.keyring follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:07:47 localhost python3[57830]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230866.5597894-91943-273270657953547/source dest=/var/lib/tripleo-config/ceph/ceph.client.manila.keyring mode=600 _original_basename=ceph.client.manila.keyring follow=False checksum=f2d03960346e263c01a33909a2839e910d075054 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:47 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.10 scrub starts Nov 27 03:07:47 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.10 scrub ok Nov 27 03:07:47 localhost ceph-osd[32722]: osd.3 pg_epoch: 48 pg[7.4( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48) [2,3,4] r=1 lpr=48 pi=[40,48)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:47 localhost ceph-osd[32722]: osd.3 pg_epoch: 48 pg[7.c( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=48) [2,3,4] r=1 lpr=48 pi=[40,48)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:07:50 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.16 scrub starts Nov 27 03:07:50 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.16 scrub ok Nov 27 03:07:51 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.1 scrub starts Nov 27 03:07:51 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.1 scrub ok Nov 27 03:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 1200.1 total, 600.0 interval#012Cumulative writes: 4289 writes, 20K keys, 4289 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4289 writes, 303 syncs, 14.16 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 1040 writes, 4028 keys, 1040 commit groups, 1.0 writes per commit group, ingest: 1.64 MB, 0.00 MB/s#012Interval WAL: 1040 writes, 163 syncs, 6.38 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 6.5e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 6.5e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.1 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 m Nov 27 03:07:52 localhost python3[57892]: ansible-ansible.legacy.stat Invoked with path=/var/lib/tripleo-config/ceph/ceph.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:07:52 localhost python3[57935]: ansible-ansible.legacy.copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230872.1494-91943-80586258653985/source dest=/var/lib/tripleo-config/ceph/ceph.conf mode=644 _original_basename=ceph.conf follow=False checksum=2b7580b6a3a12b4dfd73c9fe0a8c562335b888b1 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:53 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.d scrub starts Nov 27 03:07:53 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.d scrub ok Nov 27 03:07:54 localhost ceph-osd[31760]: osd.0 pg_epoch: 50 pg[7.d( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=50 pruub=9.001027107s) [4,5,0] r=2 lpr=50 pi=[40,50)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1211.822631836s@ mbc={}] start_peering_interval up [0,1,5] -> [4,5,0], acting [0,1,5] -> [4,5,0], acting_primary 0 -> 4, up_primary 0 -> 4, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:54 localhost ceph-osd[31760]: osd.0 pg_epoch: 50 pg[7.5( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=50 pruub=9.000842094s) [4,5,0] r=2 lpr=50 pi=[40,50)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1211.822631836s@ mbc={}] start_peering_interval up [0,1,5] -> [4,5,0], acting [0,1,5] -> [4,5,0], acting_primary 0 -> 4, up_primary 0 -> 4, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:54 localhost ceph-osd[31760]: osd.0 pg_epoch: 50 pg[7.d( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=50 pruub=9.000625610s) [4,5,0] r=2 lpr=50 pi=[40,50)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1211.822631836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:54 localhost ceph-osd[31760]: osd.0 pg_epoch: 50 pg[7.5( v 33'39 (0'0,33'39] local-lis/les=40/41 n=2 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=50 pruub=9.000401497s) [4,5,0] r=2 lpr=50 pi=[40,50)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1211.822631836s@ mbc={}] state: transitioning to Stray Nov 27 03:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 1200.2 total, 600.0 interval#012Cumulative writes: 4532 writes, 21K keys, 4532 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4532 writes, 358 syncs, 12.66 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 1148 writes, 4488 keys, 1148 commit groups, 1.0 writes per commit group, ingest: 1.83 MB, 0.00 MB/s#012Interval WAL: 1148 writes, 163 syncs, 7.04 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 5.2e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 5.2e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.2 total, 600.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 m Nov 27 03:07:56 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.9 scrub starts Nov 27 03:07:56 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 7.9 scrub ok Nov 27 03:07:56 localhost ceph-osd[31760]: osd.0 pg_epoch: 52 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=44/45 n=1 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=52 pruub=13.053000450s) [5,0,4] r=1 lpr=52 pi=[44,52)/1 luod=0'0 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1218.212158203s@ mbc={}] start_peering_interval up [2,1,0] -> [5,0,4], acting [2,1,0] -> [5,0,4], acting_primary 2 -> 5, up_primary 2 -> 5, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:56 localhost ceph-osd[31760]: osd.0 pg_epoch: 52 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=44/45 n=1 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=52 pruub=13.052901268s) [5,0,4] r=1 lpr=52 pi=[44,52)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1218.212158203s@ mbc={}] state: transitioning to Stray Nov 27 03:07:56 localhost ceph-osd[31760]: osd.0 pg_epoch: 52 pg[7.6( v 33'39 (0'0,33'39] local-lis/les=44/45 n=2 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=52 pruub=13.052772522s) [5,0,4] r=1 lpr=52 pi=[44,52)/1 luod=0'0 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1218.212280273s@ mbc={}] start_peering_interval up [2,1,0] -> [5,0,4], acting [2,1,0] -> [5,0,4], acting_primary 2 -> 5, up_primary 2 -> 5, role 2 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:56 localhost ceph-osd[31760]: osd.0 pg_epoch: 52 pg[7.6( v 33'39 (0'0,33'39] local-lis/les=44/45 n=2 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=52 pruub=13.052638054s) [5,0,4] r=1 lpr=52 pi=[44,52)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1218.212280273s@ mbc={}] state: transitioning to Stray Nov 27 03:07:57 localhost python3[57997]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/config_step.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:07:57 localhost python3[58042]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/config_step.json force=True mode=0600 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230876.8298302-92543-280919498298345/source _original_basename=tmp_chfqk8q follow=False checksum=f17091ee142621a3c8290c8c96b5b52d67b3a864 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:58 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.13 deep-scrub starts Nov 27 03:07:58 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.13 deep-scrub ok Nov 27 03:07:58 localhost python3[58104]: ansible-ansible.legacy.stat Invoked with path=/usr/local/sbin/containers-tmpwatch follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:07:58 localhost ceph-osd[32722]: osd.3 pg_epoch: 54 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=46/47 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54 pruub=10.629062653s) [3,1,5] r=0 lpr=54 pi=[46,54)/1 luod=0'0 crt=33'39 mlcod 0'0 active pruub 1213.380371094s@ mbc={}] start_peering_interval up [5,3,4] -> [3,1,5], acting [5,3,4] -> [3,1,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:58 localhost ceph-osd[32722]: osd.3 pg_epoch: 54 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=46/47 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54 pruub=10.629062653s) [3,1,5] r=0 lpr=54 pi=[46,54)/1 crt=33'39 mlcod 0'0 unknown pruub 1213.380371094s@ mbc={}] state: transitioning to Primary Nov 27 03:07:58 localhost ceph-osd[32722]: osd.3 pg_epoch: 54 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=46/47 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54 pruub=10.628812790s) [3,1,5] r=0 lpr=54 pi=[46,54)/1 luod=0'0 crt=33'39 mlcod 0'0 active pruub 1213.380615234s@ mbc={}] start_peering_interval up [5,3,4] -> [3,1,5], acting [5,3,4] -> [3,1,5], acting_primary 5 -> 3, up_primary 5 -> 3, role 1 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:07:58 localhost ceph-osd[32722]: osd.3 pg_epoch: 54 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=46/47 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54 pruub=10.628812790s) [3,1,5] r=0 lpr=54 pi=[46,54)/1 crt=33'39 mlcod 0'0 unknown pruub 1213.380615234s@ mbc={}] state: transitioning to Primary Nov 27 03:07:59 localhost python3[58147]: ansible-ansible.legacy.copy Invoked with dest=/usr/local/sbin/containers-tmpwatch group=root mode=493 owner=root src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230878.4113631-92629-216930569674501/source _original_basename=tmp_3tr3_co follow=False checksum=84397b037dad9813fed388c4bcdd4871f384cd22 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:07:59 localhost python3[58177]: ansible-cron Invoked with job=/usr/local/sbin/containers-tmpwatch name=Remove old logs special_time=daily user=root state=present backup=False minute=* hour=* day=* month=* weekday=* disabled=False env=False cron_file=None insertafter=None insertbefore=None Nov 27 03:07:59 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.19 scrub starts Nov 27 03:07:59 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.19 scrub ok Nov 27 03:07:59 localhost ceph-osd[32722]: osd.3 pg_epoch: 55 pg[7.7( v 33'39 (0'0,33'39] local-lis/les=54/55 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54) [3,1,5] r=0 lpr=54 pi=[46,54)/1 crt=33'39 mlcod 0'0 active+degraded mbc={255={(2+1)=1}}] state: react AllReplicasActivated Activating complete Nov 27 03:07:59 localhost ceph-osd[32722]: osd.3 pg_epoch: 55 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=54/55 n=1 ec=40/31 lis/c=46/46 les/c/f=47/47/0 sis=54) [3,1,5] r=0 lpr=54 pi=[46,54)/1 crt=33'39 mlcod 0'0 active+degraded mbc={255={(2+1)=3}}] state: react AllReplicasActivated Activating complete Nov 27 03:07:59 localhost python3[58195]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_2 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:08:00 localhost ceph-osd[31760]: osd.0 pg_epoch: 56 pg[7.8( v 33'39 (0'0,33'39] local-lis/les=40/41 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=56 pruub=10.514185905s) [2,0,1] r=1 lpr=56 pi=[40,56)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1219.822265625s@ mbc={}] start_peering_interval up [0,1,5] -> [2,0,1], acting [0,1,5] -> [2,0,1], acting_primary 0 -> 2, up_primary 0 -> 2, role 0 -> 1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:00 localhost ceph-osd[31760]: osd.0 pg_epoch: 56 pg[7.8( v 33'39 (0'0,33'39] local-lis/les=40/41 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=56 pruub=10.514086723s) [2,0,1] r=1 lpr=56 pi=[40,56)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1219.822265625s@ mbc={}] state: transitioning to Stray Nov 27 03:08:01 localhost ansible-async_wrapper.py[58367]: Invoked with 90378814963 3600 /home/tripleo-admin/.ansible/tmp/ansible-tmp-1764230881.0901885-92782-227061914522426/AnsiballZ_command.py _ Nov 27 03:08:01 localhost ansible-async_wrapper.py[58370]: Starting module and watcher Nov 27 03:08:01 localhost ansible-async_wrapper.py[58370]: Start watching 58371 (3600) Nov 27 03:08:01 localhost ansible-async_wrapper.py[58371]: Start module (58371) Nov 27 03:08:01 localhost ansible-async_wrapper.py[58367]: Return async_wrapper task started. Nov 27 03:08:02 localhost python3[58391]: ansible-ansible.legacy.async_status Invoked with jid=90378814963.58367 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:08:03 localhost ceph-osd[31760]: osd.0 pg_epoch: 58 pg[7.9( v 33'39 (0'0,33'39] local-lis/les=40/41 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=58 pruub=8.480168343s) [5,4,3] r=-1 lpr=58 pi=[40,58)/1 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1219.822631836s@ mbc={}] start_peering_interval up [0,1,5] -> [5,4,3], acting [0,1,5] -> [5,4,3], acting_primary 0 -> 5, up_primary 0 -> 5, role 0 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:03 localhost ceph-osd[31760]: osd.0 pg_epoch: 58 pg[7.9( v 33'39 (0'0,33'39] local-lis/les=40/41 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=58 pruub=8.479707718s) [5,4,3] r=-1 lpr=58 pi=[40,58)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1219.822631836s@ mbc={}] state: transitioning to Stray Nov 27 03:08:03 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.14 scrub starts Nov 27 03:08:03 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.14 scrub ok Nov 27 03:08:04 localhost ceph-osd[32722]: osd.3 pg_epoch: 58 pg[7.9( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=40/40 les/c/f=41/41/0 sis=58) [5,4,3] r=2 lpr=58 pi=[40,58)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:08:04 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.14 scrub starts Nov 27 03:08:04 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.14 scrub ok Nov 27 03:08:05 localhost puppet-user[58390]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:08:05 localhost puppet-user[58390]: (file: /etc/puppet/hiera.yaml) Nov 27 03:08:05 localhost puppet-user[58390]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:08:05 localhost puppet-user[58390]: (file & line not available) Nov 27 03:08:05 localhost puppet-user[58390]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:08:05 localhost puppet-user[58390]: (file & line not available) Nov 27 03:08:05 localhost puppet-user[58390]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/database/mysql/client.pp, line: 89, column: 8) Nov 27 03:08:05 localhost puppet-user[58390]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/packages.pp, line: 39, column: 69) Nov 27 03:08:05 localhost puppet-user[58390]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.11 seconds Nov 27 03:08:05 localhost puppet-user[58390]: Notice: Applied catalog in 0.03 seconds Nov 27 03:08:05 localhost puppet-user[58390]: Application: Nov 27 03:08:05 localhost puppet-user[58390]: Initial environment: production Nov 27 03:08:05 localhost puppet-user[58390]: Converged environment: production Nov 27 03:08:05 localhost puppet-user[58390]: Run mode: user Nov 27 03:08:05 localhost puppet-user[58390]: Changes: Nov 27 03:08:05 localhost puppet-user[58390]: Events: Nov 27 03:08:05 localhost puppet-user[58390]: Resources: Nov 27 03:08:05 localhost puppet-user[58390]: Total: 10 Nov 27 03:08:05 localhost puppet-user[58390]: Time: Nov 27 03:08:05 localhost puppet-user[58390]: Schedule: 0.00 Nov 27 03:08:05 localhost puppet-user[58390]: File: 0.00 Nov 27 03:08:05 localhost puppet-user[58390]: Exec: 0.01 Nov 27 03:08:05 localhost puppet-user[58390]: Augeas: 0.01 Nov 27 03:08:05 localhost puppet-user[58390]: Transaction evaluation: 0.03 Nov 27 03:08:05 localhost puppet-user[58390]: Catalog application: 0.03 Nov 27 03:08:05 localhost puppet-user[58390]: Config retrieval: 0.15 Nov 27 03:08:05 localhost puppet-user[58390]: Last run: 1764230885 Nov 27 03:08:05 localhost puppet-user[58390]: Filebucket: 0.00 Nov 27 03:08:05 localhost puppet-user[58390]: Total: 0.04 Nov 27 03:08:05 localhost puppet-user[58390]: Version: Nov 27 03:08:05 localhost puppet-user[58390]: Config: 1764230885 Nov 27 03:08:05 localhost puppet-user[58390]: Puppet: 7.10.0 Nov 27 03:08:05 localhost ansible-async_wrapper.py[58371]: Module complete (58371) Nov 27 03:08:06 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.9 deep-scrub starts Nov 27 03:08:06 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.9 deep-scrub ok Nov 27 03:08:06 localhost ansible-async_wrapper.py[58370]: Done in kid B. Nov 27 03:08:07 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.11 deep-scrub starts Nov 27 03:08:07 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.11 deep-scrub ok Nov 27 03:08:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:08:07 localhost systemd[1]: tmp-crun.Ai2YM2.mount: Deactivated successfully. Nov 27 03:08:07 localhost podman[58505]: 2025-11-27 08:08:07.989225952 +0000 UTC m=+0.084452660 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, io.openshift.expose-services=, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, vcs-type=git, config_id=tripleo_step1, tcib_managed=true, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, build-date=2025-11-18T22:49:46Z) Nov 27 03:08:08 localhost podman[58505]: 2025-11-27 08:08:08.214989625 +0000 UTC m=+0.310216283 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, architecture=x86_64, url=https://www.redhat.com, release=1761123044, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:08:08 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:08:08 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.2 scrub starts Nov 27 03:08:08 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.2 scrub ok Nov 27 03:08:11 localhost ceph-osd[31760]: osd.0 pg_epoch: 60 pg[7.a( v 33'39 (0'0,33'39] local-lis/les=44/45 n=1 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=60 pruub=14.873229980s) [4,2,3] r=-1 lpr=60 pi=[44,60)/1 luod=0'0 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1234.212280273s@ mbc={}] start_peering_interval up [2,1,0] -> [4,2,3], acting [2,1,0] -> [4,2,3], acting_primary 2 -> 4, up_primary 2 -> 4, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:11 localhost ceph-osd[31760]: osd.0 pg_epoch: 60 pg[7.a( v 33'39 (0'0,33'39] local-lis/les=44/45 n=1 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=60 pruub=14.873112679s) [4,2,3] r=-1 lpr=60 pi=[44,60)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1234.212280273s@ mbc={}] state: transitioning to Stray Nov 27 03:08:12 localhost ceph-osd[32722]: osd.3 pg_epoch: 60 pg[7.a( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=44/44 les/c/f=45/45/0 sis=60) [4,2,3] r=2 lpr=60 pi=[44,60)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:08:12 localhost python3[58674]: ansible-ansible.legacy.async_status Invoked with jid=90378814963.58367 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:08:13 localhost python3[58690]: ansible-file Invoked with path=/var/lib/container-puppet/puppetlabs state=directory setype=svirt_sandbox_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:08:13 localhost python3[58706]: ansible-stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:08:13 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.f scrub starts Nov 27 03:08:13 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.f scrub ok Nov 27 03:08:14 localhost python3[58756]: ansible-ansible.legacy.stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:14 localhost ceph-osd[32722]: osd.3 pg_epoch: 63 pg[7.c( v 33'39 (0'0,33'39] local-lis/les=48/49 n=1 ec=40/31 lis/c=48/48 les/c/f=49/49/0 sis=63 pruub=14.405332565s) [0,5,4] r=-1 lpr=63 pi=[48,63)/1 luod=0'0 crt=33'39 mlcod 0'0 active pruub 1232.510620117s@ mbc={}] start_peering_interval up [2,3,4] -> [0,5,4], acting [2,3,4] -> [0,5,4], acting_primary 2 -> 0, up_primary 2 -> 0, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:14 localhost ceph-osd[32722]: osd.3 pg_epoch: 63 pg[7.c( v 33'39 (0'0,33'39] local-lis/les=48/49 n=1 ec=40/31 lis/c=48/48 les/c/f=49/49/0 sis=63 pruub=14.405247688s) [0,5,4] r=-1 lpr=63 pi=[48,63)/1 crt=33'39 mlcod 0'0 unknown NOTIFY pruub 1232.510620117s@ mbc={}] state: transitioning to Stray Nov 27 03:08:14 localhost ceph-osd[31760]: osd.0 pg_epoch: 63 pg[7.c( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=48/48 les/c/f=49/49/0 sis=63) [0,5,4] r=0 lpr=63 pi=[48,63)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:08:14 localhost python3[58774]: ansible-ansible.legacy.file Invoked with setype=svirt_sandbox_file_t selevel=s0 dest=/var/lib/container-puppet/puppetlabs/facter.conf _original_basename=tmphqh86n5h recurse=False state=file path=/var/lib/container-puppet/puppetlabs/facter.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:08:14 localhost python3[58804]: ansible-file Invoked with path=/opt/puppetlabs/facter state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:15 localhost ceph-osd[31760]: osd.0 pg_epoch: 64 pg[7.c( v 33'39 lc 33'9 (0'0,33'39] local-lis/les=63/64 n=1 ec=40/31 lis/c=48/48 les/c/f=49/49/0 sis=63) [0,5,4] r=0 lpr=63 pi=[48,63)/1 crt=33'39 lcod 0'0 mlcod 0'0 active+degraded m=1 mbc={255={(1+2)=1}}] state: react AllReplicasActivated Activating complete Nov 27 03:08:15 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.10 scrub starts Nov 27 03:08:15 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.10 scrub ok Nov 27 03:08:15 localhost python3[58908]: ansible-ansible.posix.synchronize Invoked with src=/opt/puppetlabs/ dest=/var/lib/container-puppet/puppetlabs/ _local_rsync_path=rsync _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] ssh_connection_multiplexing=False partial=False verify_host=False mode=push dest_port=None private_key=None recursive=None links=None perms=None times=None owner=None group=None ssh_args=None link_dest=None Nov 27 03:08:16 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.15 scrub starts Nov 27 03:08:16 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.15 scrub ok Nov 27 03:08:16 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.1b scrub starts Nov 27 03:08:16 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.1b scrub ok Nov 27 03:08:16 localhost ceph-osd[32722]: osd.3 pg_epoch: 65 pg[7.d( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=50/50 les/c/f=51/51/0 sis=65) [3,2,1] r=0 lpr=65 pi=[50,65)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Primary Nov 27 03:08:16 localhost ceph-osd[31760]: osd.0 pg_epoch: 65 pg[7.d( v 33'39 (0'0,33'39] local-lis/les=50/51 n=2 ec=40/31 lis/c=50/50 les/c/f=51/51/0 sis=65 pruub=10.642235756s) [3,2,1] r=-1 lpr=65 pi=[50,65)/1 luod=0'0 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1235.828857422s@ mbc={}] start_peering_interval up [4,5,0] -> [3,2,1], acting [4,5,0] -> [3,2,1], acting_primary 4 -> 3, up_primary 4 -> 3, role 2 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:16 localhost ceph-osd[31760]: osd.0 pg_epoch: 65 pg[7.d( v 33'39 (0'0,33'39] local-lis/les=50/51 n=2 ec=40/31 lis/c=50/50 les/c/f=51/51/0 sis=65 pruub=10.641869545s) [3,2,1] r=-1 lpr=65 pi=[50,65)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1235.828857422s@ mbc={}] state: transitioning to Stray Nov 27 03:08:16 localhost python3[58927]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:17 localhost python3[58959]: ansible-stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:08:17 localhost ceph-osd[32722]: osd.3 pg_epoch: 66 pg[7.d( v 33'39 lc 33'7 (0'0,33'39] local-lis/les=65/66 n=2 ec=40/31 lis/c=50/50 les/c/f=51/51/0 sis=65) [3,2,1] r=0 lpr=65 pi=[50,65)/1 crt=33'39 lcod 0'0 mlcod 0'0 active+degraded m=2 mbc={255={(0+3)=2}}] state: react AllReplicasActivated Activating complete Nov 27 03:08:18 localhost python3[59009]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-container-shutdown follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:18 localhost python3[59027]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-container-shutdown _original_basename=tripleo-container-shutdown recurse=False state=file path=/usr/libexec/tripleo-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:18 localhost ceph-osd[31760]: osd.0 pg_epoch: 67 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=52/53 n=1 ec=40/31 lis/c=52/52 les/c/f=53/53/0 sis=67 pruub=11.051353455s) [2,1,3] r=-1 lpr=67 pi=[52,67)/1 luod=0'0 crt=33'39 lcod 0'0 mlcod 0'0 active pruub 1238.265136719s@ mbc={}] start_peering_interval up [5,0,4] -> [2,1,3], acting [5,0,4] -> [2,1,3], acting_primary 5 -> 2, up_primary 5 -> 2, role 1 -> -1, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:18 localhost ceph-osd[31760]: osd.0 pg_epoch: 67 pg[7.e( v 33'39 (0'0,33'39] local-lis/les=52/53 n=1 ec=40/31 lis/c=52/52 les/c/f=53/53/0 sis=67 pruub=11.051266670s) [2,1,3] r=-1 lpr=67 pi=[52,67)/1 crt=33'39 lcod 0'0 mlcod 0'0 unknown NOTIFY pruub 1238.265136719s@ mbc={}] state: transitioning to Stray Nov 27 03:08:19 localhost python3[59089]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-start-podman-container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:19 localhost python3[59107]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-start-podman-container _original_basename=tripleo-start-podman-container recurse=False state=file path=/usr/libexec/tripleo-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:19 localhost ceph-osd[32722]: osd.3 pg_epoch: 67 pg[7.e( empty local-lis/les=0/0 n=0 ec=40/31 lis/c=52/52 les/c/f=53/53/0 sis=67) [2,1,3] r=2 lpr=67 pi=[52,67)/1 crt=0'0 mlcod 0'0 unknown mbc={}] state: transitioning to Stray Nov 27 03:08:20 localhost python3[59169]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/tripleo-container-shutdown.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:20 localhost python3[59187]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/tripleo-container-shutdown.service _original_basename=tripleo-container-shutdown-service recurse=False state=file path=/usr/lib/systemd/system/tripleo-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:20 localhost ceph-osd[32722]: osd.3 pg_epoch: 69 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=54/55 n=1 ec=40/31 lis/c=54/54 les/c/f=55/55/0 sis=69 pruub=11.033105850s) [2,4,3] r=2 lpr=69 pi=[54,69)/1 crt=33'39 mlcod 0'0 active pruub 1235.808105469s@ mbc={255={}}] start_peering_interval up [3,1,5] -> [2,4,3], acting [3,1,5] -> [2,4,3], acting_primary 3 -> 2, up_primary 3 -> 2, role 0 -> 2, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:20 localhost ceph-osd[32722]: osd.3 pg_epoch: 69 pg[7.f( v 33'39 (0'0,33'39] local-lis/les=54/55 n=1 ec=40/31 lis/c=54/54 les/c/f=55/55/0 sis=69 pruub=11.032985687s) [2,4,3] r=2 lpr=69 pi=[54,69)/1 crt=33'39 mlcod 0'0 unknown NOTIFY pruub 1235.808105469s@ mbc={}] state: transitioning to Stray Nov 27 03:08:20 localhost python3[59249]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:21 localhost python3[59267]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset _original_basename=91-tripleo-container-shutdown-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:21 localhost python3[59297]: ansible-systemd Invoked with name=tripleo-container-shutdown state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:08:21 localhost systemd[1]: Reloading. Nov 27 03:08:21 localhost systemd-sysv-generator[59326]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:08:21 localhost systemd-rc-local-generator[59321]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:08:22 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:08:22 localhost python3[59383]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/netns-placeholder.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:23 localhost python3[59401]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/usr/lib/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:23 localhost python3[59463]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:08:23 localhost python3[59481]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:23 localhost ceph-osd[32722]: osd.3 71 crush map has features 432629239337189376, adjusting msgr requires for clients Nov 27 03:08:23 localhost ceph-osd[32722]: osd.3 71 crush map has features 432629239337189376 was 288514051259245057, adjusting msgr requires for mons Nov 27 03:08:23 localhost ceph-osd[32722]: osd.3 71 crush map has features 3314933000854323200, adjusting msgr requires for osds Nov 27 03:08:23 localhost ceph-osd[32722]: osd.3 pg_epoch: 71 pg[2.7( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=42/42 les/c/f=43/43/0 sis=71 pruub=15.627839088s) [3,1,5] r=0 lpr=71 pi=[42,71)/1 crt=0'0 mlcod 0'0 active pruub 1243.459350586s@ mbc={}] start_peering_interval up [3,1,2] -> [3,1,5], acting [3,1,2] -> [3,1,5], acting_primary 3 -> 3, up_primary 3 -> 3, role 0 -> 0, features acting 4540138322906710015 upacting 4540138322906710015 Nov 27 03:08:23 localhost ceph-osd[32722]: osd.3 pg_epoch: 71 pg[2.7( empty local-lis/les=42/43 n=0 ec=36/17 lis/c=42/42 les/c/f=43/43/0 sis=71 pruub=15.627839088s) [3,1,5] r=0 lpr=71 pi=[42,71)/1 crt=0'0 mlcod 0'0 unknown pruub 1243.459350586s@ mbc={}] state: transitioning to Primary Nov 27 03:08:24 localhost python3[59511]: ansible-systemd Invoked with name=netns-placeholder state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:08:24 localhost systemd[1]: Reloading. Nov 27 03:08:24 localhost ceph-osd[31760]: osd.0 71 crush map has features 432629239337189376, adjusting msgr requires for clients Nov 27 03:08:24 localhost ceph-osd[31760]: osd.0 71 crush map has features 432629239337189376 was 288514051259245057, adjusting msgr requires for mons Nov 27 03:08:24 localhost ceph-osd[31760]: osd.0 71 crush map has features 3314933000854323200, adjusting msgr requires for osds Nov 27 03:08:24 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.16 scrub starts Nov 27 03:08:24 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.16 scrub ok Nov 27 03:08:24 localhost systemd-sysv-generator[59540]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:08:24 localhost systemd-rc-local-generator[59537]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:08:24 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:08:24 localhost systemd[1]: Starting Create netns directory... Nov 27 03:08:24 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 03:08:24 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 03:08:24 localhost systemd[1]: Finished Create netns directory. Nov 27 03:08:25 localhost ceph-osd[32722]: osd.3 pg_epoch: 72 pg[2.7( empty local-lis/les=71/72 n=0 ec=36/17 lis/c=42/42 les/c/f=43/43/0 sis=71) [3,1,5] r=0 lpr=71 pi=[42,71)/1 crt=0'0 mlcod 0'0 active mbc={}] state: react AllReplicasActivated Activating complete Nov 27 03:08:25 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.1f deep-scrub starts Nov 27 03:08:25 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.1f deep-scrub ok Nov 27 03:08:25 localhost python3[59568]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 03:08:26 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.a scrub starts Nov 27 03:08:26 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.a scrub ok Nov 27 03:08:27 localhost python3[59624]: ansible-tripleo_container_manage Invoked with config_id=tripleo_step2 config_dir=/var/lib/tripleo-config/container-startup-config/step_2 config_patterns=*.json config_overrides={} concurrency=5 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:08:27 localhost podman[59696]: 2025-11-27 08:08:27.339392178 +0000 UTC m=+0.101238393 container create f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud_init_logs, description=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., container_name=nova_virtqemud_init_logs, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}, com.redhat.component=openstack-nova-libvirt-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, batch=17.1_20251118.1, build-date=2025-11-19T00:35:22Z, name=rhosp17/openstack-nova-libvirt, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step2, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:08:27 localhost podman[59705]: 2025-11-27 08:08:27.369280454 +0000 UTC m=+0.110561484 container create 8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute_init_log, batch=17.1_20251118.1, config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, container_name=nova_compute_init_log, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, tcib_managed=true, io.buildah.version=1.41.4, config_id=tripleo_step2, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, version=17.1.12, com.redhat.component=openstack-nova-compute-container, architecture=x86_64) Nov 27 03:08:27 localhost podman[59696]: 2025-11-27 08:08:27.287136308 +0000 UTC m=+0.048982543 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:08:27 localhost systemd[1]: Started libpod-conmon-f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b.scope. Nov 27 03:08:27 localhost systemd[1]: Started libpod-conmon-8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59.scope. Nov 27 03:08:27 localhost podman[59705]: 2025-11-27 08:08:27.304465185 +0000 UTC m=+0.045746245 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:08:27 localhost systemd[1]: Started libcrun container. Nov 27 03:08:27 localhost systemd[1]: Started libcrun container. Nov 27 03:08:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/9c4a2d4bbd63eee01920b667bdbabffc28c525da9df74ba64643b157be520c27/merged/var/log/swtpm supports timestamps until 2038 (0x7fffffff) Nov 27 03:08:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/98d9a2882c649e01bed5acc15c2ed1bb99da7cbda7e3a9702d65eb76d489868d/merged/var/log/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:08:27 localhost podman[59705]: 2025-11-27 08:08:27.430141217 +0000 UTC m=+0.171422227 container init 8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute_init_log, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, tcib_managed=true, release=1761123044, io.buildah.version=1.41.4, config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute_init_log, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, config_id=tripleo_step2, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=) Nov 27 03:08:27 localhost podman[59705]: 2025-11-27 08:08:27.437934367 +0000 UTC m=+0.179215397 container start 8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute_init_log, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute_init_log, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_id=tripleo_step2, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, release=1761123044, build-date=2025-11-19T00:36:58Z, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:08:27 localhost python3[59624]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_compute_init_log --conmon-pidfile /run/nova_compute_init_log.pid --detach=True --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --label config_id=tripleo_step2 --label container_name=nova_compute_init_log --label managed_by=tripleo_ansible --label config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_compute_init_log.log --network none --privileged=False --user root --volume /var/log/containers/nova:/var/log/nova:z registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 /bin/bash -c chown -R nova:nova /var/log/nova Nov 27 03:08:27 localhost systemd[1]: libpod-8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59.scope: Deactivated successfully. Nov 27 03:08:27 localhost podman[59696]: 2025-11-27 08:08:27.48582264 +0000 UTC m=+0.247668835 container init f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud_init_logs, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, io.openshift.expose-services=, build-date=2025-11-19T00:35:22Z, com.redhat.component=openstack-nova-libvirt-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step2, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, vendor=Red Hat, Inc., managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=nova_virtqemud_init_logs, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:08:27 localhost podman[59696]: 2025-11-27 08:08:27.496904919 +0000 UTC m=+0.258751104 container start f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud_init_logs, architecture=x86_64, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, name=rhosp17/openstack-nova-libvirt, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., distribution-scope=public, config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}, config_id=tripleo_step2, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-libvirt-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_virtqemud_init_logs, io.openshift.expose-services=, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, maintainer=OpenStack TripleO Team, version=17.1.12) Nov 27 03:08:27 localhost python3[59624]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtqemud_init_logs --conmon-pidfile /run/nova_virtqemud_init_logs.pid --detach=True --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --label config_id=tripleo_step2 --label container_name=nova_virtqemud_init_logs --label managed_by=tripleo_ansible --label config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtqemud_init_logs.log --network none --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --user root --volume /var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 /bin/bash -c chown -R tss:tss /var/log/swtpm Nov 27 03:08:27 localhost systemd[1]: libpod-f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b.scope: Deactivated successfully. Nov 27 03:08:27 localhost podman[59736]: 2025-11-27 08:08:27.55367734 +0000 UTC m=+0.096975558 container died 8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute_init_log, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., release=1761123044, config_id=tripleo_step2, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-19T00:36:58Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute_init_log, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:08:27 localhost podman[59737]: 2025-11-27 08:08:27.63886431 +0000 UTC m=+0.180675287 container cleanup 8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute_init_log, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., managed_by=tripleo_ansible, container_name=nova_compute_init_log, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'command': ['/bin/bash', '-c', 'chown -R nova:nova /var/log/nova'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'user': 'root', 'volumes': ['/var/log/containers/nova:/var/log/nova:z']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, distribution-scope=public, build-date=2025-11-19T00:36:58Z, vcs-type=git, config_id=tripleo_step2, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044) Nov 27 03:08:27 localhost systemd[1]: libpod-conmon-8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59.scope: Deactivated successfully. Nov 27 03:08:27 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.d scrub starts Nov 27 03:08:27 localhost podman[59763]: 2025-11-27 08:08:27.675360455 +0000 UTC m=+0.161573562 container died f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud_init_logs, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-nova-libvirt-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, managed_by=tripleo_ansible, release=1761123044, container_name=nova_virtqemud_init_logs, build-date=2025-11-19T00:35:22Z, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step2, config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-libvirt) Nov 27 03:08:27 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.d scrub ok Nov 27 03:08:27 localhost podman[59765]: 2025-11-27 08:08:27.802418003 +0000 UTC m=+0.281958070 container cleanup f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud_init_logs, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_virtqemud_init_logs, config_id=tripleo_step2, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, config_data={'command': ['/bin/bash', '-c', 'chown -R tss:tss /var/log/swtpm'], 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'none', 'privileged': True, 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'user': 'root', 'volumes': ['/var/log/containers/libvirt/swtpm:/var/log/swtpm:shared,z']}, com.redhat.component=openstack-nova-libvirt-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, batch=17.1_20251118.1) Nov 27 03:08:27 localhost systemd[1]: libpod-conmon-f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b.scope: Deactivated successfully. Nov 27 03:08:28 localhost podman[59886]: 2025-11-27 08:08:28.031901216 +0000 UTC m=+0.076003861 container create 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, tcib_managed=true, container_name=create_haproxy_wrapper, config_id=tripleo_step2, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:08:28 localhost podman[59893]: 2025-11-27 08:08:28.056033138 +0000 UTC m=+0.089795865 container create fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=create_virtlogd_wrapper, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, com.redhat.component=openstack-nova-libvirt-container, build-date=2025-11-19T00:35:22Z, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, distribution-scope=public, url=https://www.redhat.com, managed_by=tripleo_ansible, io.buildah.version=1.41.4, config_id=tripleo_step2, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, tcib_managed=true, vcs-type=git) Nov 27 03:08:28 localhost systemd[1]: Started libpod-conmon-5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4.scope. Nov 27 03:08:28 localhost systemd[1]: Started libpod-conmon-fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb.scope. Nov 27 03:08:28 localhost podman[59886]: 2025-11-27 08:08:27.993336105 +0000 UTC m=+0.037438750 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 03:08:28 localhost systemd[1]: Started libcrun container. Nov 27 03:08:28 localhost systemd[1]: Started libcrun container. Nov 27 03:08:28 localhost podman[59893]: 2025-11-27 08:08:27.999457351 +0000 UTC m=+0.033220108 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:08:28 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/afe1964eebc2f98c94a644ea3dcabf107e5de6e9ec86f6da43c64be956cfdb91/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 03:08:28 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6a2ac0216c924e431b841571d5cba002956cf6f55a4989caaf58ecd90b99854a/merged/var/lib/container-config-scripts supports timestamps until 2038 (0x7fffffff) Nov 27 03:08:28 localhost podman[59886]: 2025-11-27 08:08:28.1094551 +0000 UTC m=+0.153591096 container init 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, build-date=2025-11-19T00:14:25Z, architecture=x86_64, vendor=Red Hat, Inc., config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=create_haproxy_wrapper, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step2, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12) Nov 27 03:08:28 localhost podman[59893]: 2025-11-27 08:08:28.111650798 +0000 UTC m=+0.145413525 container init fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, config_id=tripleo_step2, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, com.redhat.component=openstack-nova-libvirt-container, batch=17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:35:22Z, managed_by=tripleo_ansible, container_name=create_virtlogd_wrapper, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, io.buildah.version=1.41.4, release=1761123044, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt) Nov 27 03:08:28 localhost podman[59886]: 2025-11-27 08:08:28.119388477 +0000 UTC m=+0.163491102 container start 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., build-date=2025-11-19T00:14:25Z, architecture=x86_64, tcib_managed=true, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step2, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=create_haproxy_wrapper, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4) Nov 27 03:08:28 localhost podman[59886]: 2025-11-27 08:08:28.119598063 +0000 UTC m=+0.163700688 container attach 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, tcib_managed=true, io.buildah.version=1.41.4, config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=create_haproxy_wrapper, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step2, batch=17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, url=https://www.redhat.com, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:08:28 localhost podman[59893]: 2025-11-27 08:08:28.170568208 +0000 UTC m=+0.204330975 container start fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-libvirt-container, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, url=https://www.redhat.com, config_id=tripleo_step2, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, io.buildah.version=1.41.4, build-date=2025-11-19T00:35:22Z, tcib_managed=true, container_name=create_virtlogd_wrapper, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team) Nov 27 03:08:28 localhost podman[59893]: 2025-11-27 08:08:28.171014051 +0000 UTC m=+0.204776788 container attach fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, version=17.1.12, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_step2, url=https://www.redhat.com, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, container_name=create_virtlogd_wrapper, vendor=Red Hat, Inc., distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, build-date=2025-11-19T00:35:22Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:08:28 localhost systemd[1]: var-lib-containers-storage-overlay-98d9a2882c649e01bed5acc15c2ed1bb99da7cbda7e3a9702d65eb76d489868d-merged.mount: Deactivated successfully. Nov 27 03:08:28 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-8b8c5daee9a08589ab4f26f45e433d3962b2f56b125e1b3d2be8b8a4c7c3cc59-userdata-shm.mount: Deactivated successfully. Nov 27 03:08:28 localhost systemd[1]: var-lib-containers-storage-overlay-9c4a2d4bbd63eee01920b667bdbabffc28c525da9df74ba64643b157be520c27-merged.mount: Deactivated successfully. Nov 27 03:08:28 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-f287e3489c940028cbc06e5ba418a90a0177c977a6e38e61b3f33da687e17f9b-userdata-shm.mount: Deactivated successfully. Nov 27 03:08:29 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.1 scrub starts Nov 27 03:08:29 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.1 scrub ok Nov 27 03:08:29 localhost ovs-vsctl[60015]: ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: database connection failed (No such file or directory) Nov 27 03:08:30 localhost systemd[1]: libpod-fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb.scope: Deactivated successfully. Nov 27 03:08:30 localhost systemd[1]: libpod-fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb.scope: Consumed 2.114s CPU time. Nov 27 03:08:30 localhost podman[59893]: 2025-11-27 08:08:30.228392182 +0000 UTC m=+2.262154929 container died fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, version=17.1.12, release=1761123044, config_id=tripleo_step2, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=create_virtlogd_wrapper, com.redhat.component=openstack-nova-libvirt-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, build-date=2025-11-19T00:35:22Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., tcib_managed=true, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, url=https://www.redhat.com, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:08:30 localhost systemd[1]: tmp-crun.eum0eR.mount: Deactivated successfully. Nov 27 03:08:30 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb-userdata-shm.mount: Deactivated successfully. Nov 27 03:08:30 localhost podman[60140]: 2025-11-27 08:08:30.330727333 +0000 UTC m=+0.092783374 container cleanup fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=create_virtlogd_wrapper, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step2, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, build-date=2025-11-19T00:35:22Z, version=17.1.12, container_name=create_virtlogd_wrapper, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']}, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, name=rhosp17/openstack-nova-libvirt, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, batch=17.1_20251118.1, com.redhat.component=openstack-nova-libvirt-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64) Nov 27 03:08:30 localhost systemd[1]: libpod-conmon-fcf399900c3f98010811a17f6bfbb6d1a5351ddbb902a407c6325154958409eb.scope: Deactivated successfully. Nov 27 03:08:30 localhost python3[59624]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name create_virtlogd_wrapper --cgroupns=host --conmon-pidfile /run/create_virtlogd_wrapper.pid --detach=False --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --label config_id=tripleo_step2 --label container_name=create_virtlogd_wrapper --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::nova::virtlogd_wrapper'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/create_virtlogd_wrapper.log --network host --pid host --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /var/lib/container-config-scripts:/var/lib/container-config-scripts:shared,z registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 /container_puppet_apply.sh 4 file include ::tripleo::profile::base::nova::virtlogd_wrapper Nov 27 03:08:31 localhost systemd[1]: libpod-5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4.scope: Deactivated successfully. Nov 27 03:08:31 localhost systemd[1]: libpod-5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4.scope: Consumed 2.041s CPU time. Nov 27 03:08:31 localhost podman[59886]: 2025-11-27 08:08:31.016926682 +0000 UTC m=+3.061029327 container died 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., tcib_managed=true, container_name=create_haproxy_wrapper, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step2, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, release=1761123044, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, version=17.1.12, distribution-scope=public) Nov 27 03:08:31 localhost podman[60178]: 2025-11-27 08:08:31.106737766 +0000 UTC m=+0.076704451 container cleanup 5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=create_haproxy_wrapper, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, distribution-scope=public, config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=create_haproxy_wrapper, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_id=tripleo_step2, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, version=17.1.12, io.openshift.expose-services=, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:08:31 localhost systemd[1]: libpod-conmon-5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4.scope: Deactivated successfully. Nov 27 03:08:31 localhost python3[59624]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name create_haproxy_wrapper --conmon-pidfile /run/create_haproxy_wrapper.pid --detach=False --label config_id=tripleo_step2 --label container_name=create_haproxy_wrapper --label managed_by=tripleo_ansible --label config_data={'command': ['/container_puppet_apply.sh', '4', 'file', 'include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers'], 'detach': False, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'start_order': 1, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/create_haproxy_wrapper.log --network host --pid host --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z --volume /var/lib/neutron:/var/lib/neutron:shared,z registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 /container_puppet_apply.sh 4 file include ::tripleo::profile::base::neutron::ovn_metadata_agent_wrappers Nov 27 03:08:31 localhost systemd[1]: var-lib-containers-storage-overlay-6a2ac0216c924e431b841571d5cba002956cf6f55a4989caaf58ecd90b99854a-merged.mount: Deactivated successfully. Nov 27 03:08:31 localhost systemd[1]: var-lib-containers-storage-overlay-afe1964eebc2f98c94a644ea3dcabf107e5de6e9ec86f6da43c64be956cfdb91-merged.mount: Deactivated successfully. Nov 27 03:08:31 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-5a6c760e98bd0b3274e078ce78ba9e9c4973f1cbba75387f79b81904a82ad8a4-userdata-shm.mount: Deactivated successfully. Nov 27 03:08:31 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.7 scrub starts Nov 27 03:08:31 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 5.7 scrub ok Nov 27 03:08:31 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.c deep-scrub starts Nov 27 03:08:31 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.c deep-scrub ok Nov 27 03:08:31 localhost python3[60232]: ansible-file Invoked with path=/var/lib/container-puppet/container-puppet-tasks2.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:33 localhost python3[60353]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=True puppet_config=/var/lib/container-puppet/container-puppet-tasks2.json short_hostname=np0005537446 step=2 update_config_hash_only=False Nov 27 03:08:33 localhost python3[60369]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:08:34 localhost python3[60385]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_2 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:08:34 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.2 scrub starts Nov 27 03:08:34 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.2 scrub ok Nov 27 03:08:38 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.3 scrub starts Nov 27 03:08:38 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.3 scrub ok Nov 27 03:08:38 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.3 scrub starts Nov 27 03:08:38 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.3 scrub ok Nov 27 03:08:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:08:38 localhost podman[60386]: 2025-11-27 08:08:38.993264455 +0000 UTC m=+0.082324843 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, release=1761123044, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, distribution-scope=public) Nov 27 03:08:39 localhost podman[60386]: 2025-11-27 08:08:39.182910083 +0000 UTC m=+0.271970481 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, container_name=metrics_qdr, release=1761123044, tcib_managed=true, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vendor=Red Hat, Inc., com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, name=rhosp17/openstack-qdrouterd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, maintainer=OpenStack TripleO Team) Nov 27 03:08:39 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:08:40 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.1a scrub starts Nov 27 03:08:40 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.1a scrub ok Nov 27 03:08:40 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.5 scrub starts Nov 27 03:08:40 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.5 scrub ok Nov 27 03:08:44 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.1c scrub starts Nov 27 03:08:44 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.1c scrub ok Nov 27 03:08:46 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.5 scrub starts Nov 27 03:08:46 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.5 scrub ok Nov 27 03:08:46 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.a scrub starts Nov 27 03:08:46 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 4.a scrub ok Nov 27 03:08:47 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.1a deep-scrub starts Nov 27 03:08:47 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.1a deep-scrub ok Nov 27 03:08:47 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.a scrub starts Nov 27 03:08:47 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 6.a scrub ok Nov 27 03:08:49 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.18 scrub starts Nov 27 03:08:49 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.18 scrub ok Nov 27 03:08:51 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.1b scrub starts Nov 27 03:08:51 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.1b scrub ok Nov 27 03:08:52 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.f scrub starts Nov 27 03:08:52 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.1a deep-scrub starts Nov 27 03:08:52 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.f scrub ok Nov 27 03:08:56 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.16 scrub starts Nov 27 03:08:56 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.16 scrub ok Nov 27 03:08:57 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.e scrub starts Nov 27 03:08:57 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.e scrub ok Nov 27 03:08:59 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.5 scrub starts Nov 27 03:08:59 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.5 scrub ok Nov 27 03:09:01 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.2 scrub starts Nov 27 03:09:01 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.2 scrub ok Nov 27 03:09:02 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.7 scrub starts Nov 27 03:09:02 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.7 scrub ok Nov 27 03:09:05 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.d deep-scrub starts Nov 27 03:09:05 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.d deep-scrub ok Nov 27 03:09:08 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.c scrub starts Nov 27 03:09:08 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.c scrub ok Nov 27 03:09:08 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.e deep-scrub starts Nov 27 03:09:08 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.e deep-scrub ok Nov 27 03:09:09 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.8 scrub starts Nov 27 03:09:09 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.8 scrub ok Nov 27 03:09:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:09:09 localhost systemd[1]: tmp-crun.gbgcMC.mount: Deactivated successfully. Nov 27 03:09:09 localhost podman[60415]: 2025-11-27 08:09:09.99000499 +0000 UTC m=+0.082768304 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_id=tripleo_step1, build-date=2025-11-18T22:49:46Z, tcib_managed=true, container_name=metrics_qdr, io.openshift.expose-services=, vcs-type=git, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, name=rhosp17/openstack-qdrouterd, release=1761123044, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:09:10 localhost podman[60415]: 2025-11-27 08:09:10.201082475 +0000 UTC m=+0.293845819 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, url=https://www.redhat.com, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.openshift.expose-services=, vendor=Red Hat, Inc., container_name=metrics_qdr, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:09:10 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:09:11 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.8 scrub starts Nov 27 03:09:11 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.8 scrub ok Nov 27 03:09:13 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.18 scrub starts Nov 27 03:09:13 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.15 scrub starts Nov 27 03:09:13 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 5.18 scrub ok Nov 27 03:09:14 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.3 scrub starts Nov 27 03:09:14 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 3.3 scrub ok Nov 27 03:09:15 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.11 scrub starts Nov 27 03:09:15 localhost ceph-osd[31760]: log_channel(cluster) log [DBG] : 2.11 scrub ok Nov 27 03:09:20 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 7.7 deep-scrub starts Nov 27 03:09:20 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 7.7 deep-scrub ok Nov 27 03:09:21 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.7 scrub starts Nov 27 03:09:21 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.7 scrub ok Nov 27 03:09:27 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.0 deep-scrub starts Nov 27 03:09:27 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.0 deep-scrub ok Nov 27 03:09:29 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.13 scrub starts Nov 27 03:09:29 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.13 scrub ok Nov 27 03:09:30 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.d scrub starts Nov 27 03:09:30 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 3.d scrub ok Nov 27 03:09:36 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.17 scrub starts Nov 27 03:09:36 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 2.17 scrub ok Nov 27 03:09:37 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.1a scrub starts Nov 27 03:09:37 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 4.1a scrub ok Nov 27 03:09:38 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.15 deep-scrub starts Nov 27 03:09:38 localhost ceph-osd[32722]: log_channel(cluster) log [DBG] : 6.15 deep-scrub ok Nov 27 03:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:09:40 localhost podman[60522]: 2025-11-27 08:09:40.977821718 +0000 UTC m=+0.068740774 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, config_id=tripleo_step1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, tcib_managed=true, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, distribution-scope=public, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git) Nov 27 03:09:41 localhost podman[60522]: 2025-11-27 08:09:41.167899102 +0000 UTC m=+0.258818138 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, release=1761123044, version=17.1.12, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, architecture=x86_64) Nov 27 03:09:41 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:10:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:10:12 localhost podman[60552]: 2025-11-27 08:10:12.021924734 +0000 UTC m=+0.099337092 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, config_id=tripleo_step1, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, url=https://www.redhat.com, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, release=1761123044) Nov 27 03:10:12 localhost podman[60552]: 2025-11-27 08:10:12.222468087 +0000 UTC m=+0.299880455 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, io.openshift.expose-services=, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, batch=17.1_20251118.1, release=1761123044, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, vcs-type=git, url=https://www.redhat.com, distribution-scope=public, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 03:10:12 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:10:12 localhost systemd[1]: tmp-crun.cLYmgY.mount: Deactivated successfully. Nov 27 03:10:12 localhost podman[60679]: 2025-11-27 08:10:12.953822491 +0000 UTC m=+0.106486570 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, io.openshift.expose-services=, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, GIT_BRANCH=main, com.redhat.component=rhceph-container, name=rhceph, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_CLEAN=True) Nov 27 03:10:13 localhost podman[60679]: 2025-11-27 08:10:13.057197323 +0000 UTC m=+0.209861432 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, version=7, io.openshift.expose-services=, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , architecture=x86_64, distribution-scope=public, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7) Nov 27 03:10:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:10:42 localhost podman[60820]: 2025-11-27 08:10:42.995944688 +0000 UTC m=+0.085218870 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, architecture=x86_64, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, vcs-type=git, config_id=tripleo_step1, distribution-scope=public, container_name=metrics_qdr, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, release=1761123044, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd) Nov 27 03:10:43 localhost podman[60820]: 2025-11-27 08:10:43.193956602 +0000 UTC m=+0.283230774 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, version=17.1.12, container_name=metrics_qdr, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd) Nov 27 03:10:43 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:11:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:11:13 localhost podman[60848]: 2025-11-27 08:11:13.98123133 +0000 UTC m=+0.082039400 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, version=17.1.12, vendor=Red Hat, Inc., architecture=x86_64, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044) Nov 27 03:11:14 localhost podman[60848]: 2025-11-27 08:11:14.143973983 +0000 UTC m=+0.244782083 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, version=17.1.12, config_id=tripleo_step1, managed_by=tripleo_ansible, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:11:14 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:11:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:11:44 localhost podman[60956]: 2025-11-27 08:11:44.983047498 +0000 UTC m=+0.076845296 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, architecture=x86_64, distribution-scope=public) Nov 27 03:11:45 localhost podman[60956]: 2025-11-27 08:11:45.21496967 +0000 UTC m=+0.308767468 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, version=17.1.12, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.buildah.version=1.41.4, vendor=Red Hat, Inc., managed_by=tripleo_ansible, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1) Nov 27 03:11:45 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:12:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:12:15 localhost podman[60983]: 2025-11-27 08:12:15.979173076 +0000 UTC m=+0.074702785 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, tcib_managed=true, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, io.openshift.expose-services=, managed_by=tripleo_ansible, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1) Nov 27 03:12:16 localhost podman[60983]: 2025-11-27 08:12:16.175933412 +0000 UTC m=+0.271463081 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, container_name=metrics_qdr, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:12:16 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:12:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:12:46 localhost systemd[1]: tmp-crun.DtQ5BW.mount: Deactivated successfully. Nov 27 03:12:46 localhost podman[61087]: 2025-11-27 08:12:46.984610504 +0000 UTC m=+0.080347493 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, batch=17.1_20251118.1, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, config_id=tripleo_step1, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:12:47 localhost podman[61087]: 2025-11-27 08:12:47.15780569 +0000 UTC m=+0.253542629 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64, tcib_managed=true, config_id=tripleo_step1, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:12:47 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:13:08 localhost python3[61163]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/config_step.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:08 localhost python3[61208]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/config_step.json force=True mode=0600 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231188.2223368-98889-194151091216388/source _original_basename=tmpecmel0ru follow=False checksum=62439dd24dde40c90e7a39f6a1b31cc6061fe59b backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:10 localhost python3[61238]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_3 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:11 localhost ansible-async_wrapper.py[61410]: Invoked with 438660151732 3600 /home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231191.358843-99127-186380316064807/AnsiballZ_command.py _ Nov 27 03:13:11 localhost ansible-async_wrapper.py[61413]: Starting module and watcher Nov 27 03:13:11 localhost ansible-async_wrapper.py[61413]: Start watching 61414 (3600) Nov 27 03:13:11 localhost ansible-async_wrapper.py[61414]: Start module (61414) Nov 27 03:13:11 localhost ansible-async_wrapper.py[61410]: Return async_wrapper task started. Nov 27 03:13:12 localhost python3[61434]: ansible-ansible.legacy.async_status Invoked with jid=438660151732.61410 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:13:15 localhost puppet-user[61433]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:13:15 localhost puppet-user[61433]: (file: /etc/puppet/hiera.yaml) Nov 27 03:13:15 localhost puppet-user[61433]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:13:15 localhost puppet-user[61433]: (file & line not available) Nov 27 03:13:15 localhost puppet-user[61433]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:13:15 localhost puppet-user[61433]: (file & line not available) Nov 27 03:13:15 localhost puppet-user[61433]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/database/mysql/client.pp, line: 89, column: 8) Nov 27 03:13:15 localhost puppet-user[61433]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/packages.pp, line: 39, column: 69) Nov 27 03:13:15 localhost puppet-user[61433]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.11 seconds Nov 27 03:13:15 localhost puppet-user[61433]: Notice: Applied catalog in 0.03 seconds Nov 27 03:13:15 localhost puppet-user[61433]: Application: Nov 27 03:13:15 localhost puppet-user[61433]: Initial environment: production Nov 27 03:13:15 localhost puppet-user[61433]: Converged environment: production Nov 27 03:13:15 localhost puppet-user[61433]: Run mode: user Nov 27 03:13:15 localhost puppet-user[61433]: Changes: Nov 27 03:13:15 localhost puppet-user[61433]: Events: Nov 27 03:13:15 localhost puppet-user[61433]: Resources: Nov 27 03:13:15 localhost puppet-user[61433]: Total: 10 Nov 27 03:13:15 localhost puppet-user[61433]: Time: Nov 27 03:13:15 localhost puppet-user[61433]: Schedule: 0.00 Nov 27 03:13:15 localhost puppet-user[61433]: File: 0.00 Nov 27 03:13:15 localhost puppet-user[61433]: Exec: 0.01 Nov 27 03:13:15 localhost puppet-user[61433]: Augeas: 0.01 Nov 27 03:13:15 localhost puppet-user[61433]: Transaction evaluation: 0.03 Nov 27 03:13:15 localhost puppet-user[61433]: Catalog application: 0.03 Nov 27 03:13:15 localhost puppet-user[61433]: Config retrieval: 0.15 Nov 27 03:13:15 localhost puppet-user[61433]: Last run: 1764231195 Nov 27 03:13:15 localhost puppet-user[61433]: Filebucket: 0.00 Nov 27 03:13:15 localhost puppet-user[61433]: Total: 0.04 Nov 27 03:13:15 localhost puppet-user[61433]: Version: Nov 27 03:13:15 localhost puppet-user[61433]: Config: 1764231195 Nov 27 03:13:15 localhost puppet-user[61433]: Puppet: 7.10.0 Nov 27 03:13:15 localhost ansible-async_wrapper.py[61414]: Module complete (61414) Nov 27 03:13:16 localhost ansible-async_wrapper.py[61413]: Done in kid B. Nov 27 03:13:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:13:17 localhost podman[61545]: 2025-11-27 08:13:17.979506298 +0000 UTC m=+0.074450337 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, tcib_managed=true, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, batch=17.1_20251118.1, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd) Nov 27 03:13:18 localhost podman[61545]: 2025-11-27 08:13:18.174807003 +0000 UTC m=+0.269751012 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.expose-services=, config_id=tripleo_step1, distribution-scope=public, name=rhosp17/openstack-qdrouterd, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, url=https://www.redhat.com, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:13:18 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:13:23 localhost python3[61665]: ansible-ansible.legacy.async_status Invoked with jid=438660151732.61410 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:13:23 localhost python3[61681]: ansible-file Invoked with path=/var/lib/container-puppet/puppetlabs state=directory setype=svirt_sandbox_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:13:24 localhost python3[61697]: ansible-stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:24 localhost python3[61747]: ansible-ansible.legacy.stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:25 localhost python3[61765]: ansible-ansible.legacy.file Invoked with setype=svirt_sandbox_file_t selevel=s0 dest=/var/lib/container-puppet/puppetlabs/facter.conf _original_basename=tmp8mn0qzp2 recurse=False state=file path=/var/lib/container-puppet/puppetlabs/facter.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:13:25 localhost python3[61795]: ansible-file Invoked with path=/opt/puppetlabs/facter state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:27 localhost python3[61898]: ansible-ansible.posix.synchronize Invoked with src=/opt/puppetlabs/ dest=/var/lib/container-puppet/puppetlabs/ _local_rsync_path=rsync _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] ssh_connection_multiplexing=False partial=False verify_host=False mode=push dest_port=None private_key=None recursive=None links=None perms=None times=None owner=None group=None ssh_args=None link_dest=None Nov 27 03:13:28 localhost python3[61917]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:29 localhost python3[61949]: ansible-stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:29 localhost python3[61999]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-container-shutdown follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:30 localhost python3[62017]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-container-shutdown _original_basename=tripleo-container-shutdown recurse=False state=file path=/usr/libexec/tripleo-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:30 localhost python3[62079]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-start-podman-container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:30 localhost python3[62097]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-start-podman-container _original_basename=tripleo-start-podman-container recurse=False state=file path=/usr/libexec/tripleo-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:31 localhost python3[62159]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/tripleo-container-shutdown.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:31 localhost python3[62177]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/tripleo-container-shutdown.service _original_basename=tripleo-container-shutdown-service recurse=False state=file path=/usr/lib/systemd/system/tripleo-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:32 localhost python3[62239]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:32 localhost python3[62257]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset _original_basename=91-tripleo-container-shutdown-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:33 localhost python3[62287]: ansible-systemd Invoked with name=tripleo-container-shutdown state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:33 localhost systemd[1]: Reloading. Nov 27 03:13:33 localhost systemd-sysv-generator[62317]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:33 localhost systemd-rc-local-generator[62312]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:33 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:33 localhost python3[62373]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/netns-placeholder.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:34 localhost python3[62391]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/usr/lib/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:34 localhost python3[62453]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:13:35 localhost python3[62471]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:35 localhost python3[62501]: ansible-systemd Invoked with name=netns-placeholder state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:35 localhost systemd[1]: Reloading. Nov 27 03:13:35 localhost systemd-sysv-generator[62527]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:35 localhost systemd-rc-local-generator[62523]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:35 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:35 localhost systemd[1]: Starting Create netns directory... Nov 27 03:13:35 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 03:13:35 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 03:13:35 localhost systemd[1]: Finished Create netns directory. Nov 27 03:13:36 localhost python3[62560]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 03:13:38 localhost python3[62618]: ansible-tripleo_container_manage Invoked with config_id=tripleo_step3 config_dir=/var/lib/tripleo-config/container-startup-config/step_3 config_patterns=*.json config_overrides={} concurrency=5 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:13:39 localhost podman[62774]: 2025-11-27 08:13:39.240804329 +0000 UTC m=+0.063113041 container create 213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_init_log, url=https://www.redhat.com, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.expose-services=, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_init_log) Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b.scope. Nov 27 03:13:39 localhost podman[62794]: 2025-11-27 08:13:39.278733632 +0000 UTC m=+0.080641261 container create 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-rsyslog-container, name=rhosp17/openstack-rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 rsyslog, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 rsyslog, container_name=rsyslog, io.buildah.version=1.41.4, config_id=tripleo_step3, release=1761123044, build-date=2025-11-18T22:49:49Z, version=17.1.12) Nov 27 03:13:39 localhost systemd[1]: tmp-crun.xw3Cbl.mount: Deactivated successfully. Nov 27 03:13:39 localhost podman[62789]: 2025-11-27 08:13:39.283708701 +0000 UTC m=+0.093004058 container create 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, build-date=2025-11-18T22:51:28Z, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, container_name=collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, release=1761123044, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4) Nov 27 03:13:39 localhost podman[62800]: 2025-11-27 08:13:39.292936341 +0000 UTC m=+0.086224059 container create c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-libvirt-container, release=1761123044, vcs-type=git, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, managed_by=tripleo_ansible, container_name=nova_virtlogd_wrapper, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:35:22Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., config_id=tripleo_step3, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true) Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost podman[62774]: 2025-11-27 08:13:39.206545419 +0000 UTC m=+0.028854171 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7958ccd3e0b0511dc5f6716870f0adbf868f68deb164ea6566f556cebcde5e7f/merged/var/log/ceilometer supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope. Nov 27 03:13:39 localhost podman[62774]: 2025-11-27 08:13:39.31325235 +0000 UTC m=+0.135561142 container init 213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_init_log, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, tcib_managed=true, version=17.1.12, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_init_log, batch=17.1_20251118.1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458.scope. Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost podman[62774]: 2025-11-27 08:13:39.319807973 +0000 UTC m=+0.142116725 container start 213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_init_log, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, container_name=ceilometer_init_log, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12) Nov 27 03:13:39 localhost podman[62789]: 2025-11-27 08:13:39.222468874 +0000 UTC m=+0.031764231 image pull registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name ceilometer_init_log --conmon-pidfile /run/ceilometer_init_log.pid --detach=True --label config_id=tripleo_step3 --label container_name=ceilometer_init_log --label managed_by=tripleo_ansible --label config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/ceilometer_init_log.log --network none --user root --volume /var/log/containers/ceilometer:/var/log/ceilometer:z registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 /bin/bash -c chown -R ceilometer:ceilometer /var/log/ceilometer Nov 27 03:13:39 localhost systemd[1]: libpod-213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b.scope: Deactivated successfully. Nov 27 03:13:39 localhost podman[62794]: 2025-11-27 08:13:39.330250306 +0000 UTC m=+0.132157925 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, com.redhat.component=openstack-rsyslog-container, container_name=rsyslog, batch=17.1_20251118.1, url=https://www.redhat.com, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 rsyslog, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 rsyslog, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, build-date=2025-11-18T22:49:49Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-rsyslog, tcib_managed=true, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost podman[62794]: 2025-11-27 08:13:39.237026003 +0000 UTC m=+0.038933642 image pull registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 Nov 27 03:13:39 localhost podman[62794]: 2025-11-27 08:13:39.339972159 +0000 UTC m=+0.141879798 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, container_name=rsyslog, tcib_managed=true, url=https://www.redhat.com, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.expose-services=, com.redhat.component=openstack-rsyslog-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, name=rhosp17/openstack-rsyslog, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, description=Red Hat OpenStack Platform 17.1 rsyslog, release=1761123044, architecture=x86_64, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., distribution-scope=public) Nov 27 03:13:39 localhost podman[62800]: 2025-11-27 08:13:39.244633916 +0000 UTC m=+0.037921644 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:39 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name rsyslog --conmon-pidfile /run/rsyslog.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=fd416cb72dcd90fffe349badf5624773 --label config_id=tripleo_step3 --label container_name=rsyslog --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/rsyslog.log --network host --privileged=True --security-opt label=disable --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro --volume /var/log/containers:/var/log/containers:ro --volume /var/log/containers/rsyslog:/var/log/rsyslog:rw,z --volume /var/log:/var/log/host:ro --volume /var/lib/rsyslog.container:/var/lib/rsyslog:rw,z registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1 Nov 27 03:13:39 localhost podman[62867]: 2025-11-27 08:13:39.367628514 +0000 UTC m=+0.035448644 container died 213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_init_log, architecture=x86_64, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, config_id=tripleo_step3, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_init_log, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.378745476 +0000 UTC m=+0.163815423 container create 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, release=1761123044, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, url=https://www.redhat.com, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, container_name=nova_statedir_owner, vcs-type=git, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:13:39 localhost podman[62800]: 2025-11-27 08:13:39.390652579 +0000 UTC m=+0.183940307 container init c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, batch=17.1_20251118.1, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., version=17.1.12, architecture=x86_64, managed_by=tripleo_ansible, container_name=nova_virtlogd_wrapper, vcs-type=git, com.redhat.component=openstack-nova-libvirt-container, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:35:22Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, name=rhosp17/openstack-nova-libvirt, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}) Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5.scope. Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4cf720646dffd6348676ebc255c10bdddf478a8f172f7ca0f19f021e215e5abf/merged/container-config-scripts supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4cf720646dffd6348676ebc255c10bdddf478a8f172f7ca0f19f021e215e5abf/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4cf720646dffd6348676ebc255c10bdddf478a8f172f7ca0f19f021e215e5abf/merged/var/lib/_nova_secontext supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.419365105 +0000 UTC m=+0.204435052 container init 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, container_name=nova_statedir_owner, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, release=1761123044, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step3, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, tcib_managed=true) Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.427464042 +0000 UTC m=+0.212533989 container start 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, url=https://www.redhat.com, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, managed_by=tripleo_ansible, vcs-type=git, batch=17.1_20251118.1, container_name=nova_statedir_owner, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, release=1761123044, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, architecture=x86_64, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.427648787 +0000 UTC m=+0.212718824 container attach 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, container_name=nova_statedir_owner, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, architecture=x86_64, config_id=tripleo_step3, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:13:39 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:39 localhost systemd[1]: Created slice User Slice of UID 0. Nov 27 03:13:39 localhost systemd[1]: Starting User Runtime Directory /run/user/0... Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.344318131 +0000 UTC m=+0.129388178 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:13:39 localhost podman[62800]: 2025-11-27 08:13:39.448500111 +0000 UTC m=+0.241787829 container start c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:35:22Z, maintainer=OpenStack TripleO Team, container_name=nova_virtlogd_wrapper, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, distribution-scope=public, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.component=openstack-nova-libvirt-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, config_id=tripleo_step3, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, release=1761123044, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt) Nov 27 03:13:39 localhost systemd[1]: Finished User Runtime Directory /run/user/0. Nov 27 03:13:39 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtlogd_wrapper --cgroupns=host --conmon-pidfile /run/nova_virtlogd_wrapper.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtlogd_wrapper --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtlogd_wrapper.log --network host --pid host --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:39 localhost systemd[1]: Starting User Manager for UID 0... Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.scope. Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost systemd[1]: libpod-31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5.scope: Deactivated successfully. Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/706e7dad99f72ad0a2bbd0167294cca4c86e34b8636b5f5f9b7b80818f253353/merged/scripts supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost podman[62814]: 2025-11-27 08:13:39.47841161 +0000 UTC m=+0.263481557 container died 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, version=17.1.12, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, tcib_managed=true, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_statedir_owner, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step3, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/706e7dad99f72ad0a2bbd0167294cca4c86e34b8636b5f5f9b7b80818f253353/merged/var/log/collectd supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:13:39 localhost podman[62789]: 2025-11-27 08:13:39.501489167 +0000 UTC m=+0.310784514 container init 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.component=openstack-collectd-container, container_name=collectd, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc., tcib_managed=true, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:13:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:13:39 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:39 localhost podman[62934]: 2025-11-27 08:13:39.564060481 +0000 UTC m=+0.133334429 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, release=1761123044, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., batch=17.1_20251118.1, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, managed_by=tripleo_ansible, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:49Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, container_name=rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 rsyslog, config_id=tripleo_step3, com.redhat.component=openstack-rsyslog-container, name=rhosp17/openstack-rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:13:39 localhost podman[62934]: 2025-11-27 08:13:39.58045389 +0000 UTC m=+0.149727828 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:49Z, summary=Red Hat OpenStack Platform 17.1 rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-rsyslog-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, container_name=rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, managed_by=tripleo_ansible, io.openshift.expose-services=, name=rhosp17/openstack-rsyslog, version=17.1.12, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, maintainer=OpenStack TripleO Team, tcib_managed=true) Nov 27 03:13:39 localhost systemd[1]: libpod-conmon-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:13:39 localhost podman[62969]: 2025-11-27 08:13:39.59541812 +0000 UTC m=+0.103803171 container cleanup 31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_statedir_owner, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step3, distribution-scope=public, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, release=1761123044, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, container_name=nova_statedir_owner) Nov 27 03:13:39 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_statedir_owner --conmon-pidfile /run/nova_statedir_owner.pid --detach=False --env NOVA_STATEDIR_OWNERSHIP_SKIP=triliovault-mounts --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --env __OS_DEBUG=true --label config_id=tripleo_step3 --label container_name=nova_statedir_owner --label managed_by=tripleo_ansible --label config_data={'command': '/container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py', 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': 'triliovault-mounts', 'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719', '__OS_DEBUG': 'true'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'none', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/container-config-scripts:/container-config-scripts:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_statedir_owner.log --network none --privileged=False --security-opt label=disable --user root --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z --volume /var/lib/container-config-scripts:/container-config-scripts:z registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 /container-config-scripts/pyshim.sh /container-config-scripts/nova_statedir_ownership.py Nov 27 03:13:39 localhost systemd[62955]: Queued start job for default target Main User Target. Nov 27 03:13:39 localhost systemd[1]: libpod-conmon-31fb6d963bcb19acc13dc2350bb2fb330a69042de47155908d9cd6e560b261b5.scope: Deactivated successfully. Nov 27 03:13:39 localhost systemd[62955]: Created slice User Application Slice. Nov 27 03:13:39 localhost systemd[62955]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). Nov 27 03:13:39 localhost systemd[62955]: Started Daily Cleanup of User's Temporary Directories. Nov 27 03:13:39 localhost systemd[62955]: Reached target Paths. Nov 27 03:13:39 localhost systemd[62955]: Reached target Timers. Nov 27 03:13:39 localhost podman[62877]: 2025-11-27 08:13:39.6114917 +0000 UTC m=+0.271474101 container cleanup 213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_init_log, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, url=https://www.redhat.com, config_id=tripleo_step3, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'command': ['/bin/bash', '-c', 'chown -R ceilometer:ceilometer /var/log/ceilometer'], 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'none', 'start_order': 0, 'user': 'root', 'volumes': ['/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, tcib_managed=true, release=1761123044, container_name=ceilometer_init_log, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:13:39 localhost systemd[62955]: Starting D-Bus User Message Bus Socket... Nov 27 03:13:39 localhost systemd[62955]: Starting Create User's Volatile Files and Directories... Nov 27 03:13:39 localhost systemd[1]: libpod-conmon-213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b.scope: Deactivated successfully. Nov 27 03:13:39 localhost podman[62789]: 2025-11-27 08:13:39.626741448 +0000 UTC m=+0.436036805 container start 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, container_name=collectd, architecture=x86_64, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044) Nov 27 03:13:39 localhost systemd[62955]: Finished Create User's Volatile Files and Directories. Nov 27 03:13:39 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name collectd --cap-add IPC_LOCK --conmon-pidfile /run/collectd.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=d31718fcd17fdeee6489534105191c7a --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step3 --label container_name=collectd --label managed_by=tripleo_ansible --label config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/collectd.log --memory 512m --network host --pid host --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro --volume /var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro --volume /var/log/containers/collectd:/var/log/collectd:rw,z --volume /var/lib/container-config-scripts:/config-scripts:ro --volume /var/lib/container-user-scripts:/scripts:z --volume /run:/run:rw --volume /sys/fs/cgroup:/sys/fs/cgroup:ro registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1 Nov 27 03:13:39 localhost systemd[62955]: Listening on D-Bus User Message Bus Socket. Nov 27 03:13:39 localhost systemd[62955]: Reached target Sockets. Nov 27 03:13:39 localhost systemd[62955]: Reached target Basic System. Nov 27 03:13:39 localhost systemd[62955]: Reached target Main User Target. Nov 27 03:13:39 localhost systemd[62955]: Startup finished in 141ms. Nov 27 03:13:39 localhost systemd[1]: Started User Manager for UID 0. Nov 27 03:13:39 localhost systemd[1]: Started Session c1 of User root. Nov 27 03:13:39 localhost systemd[1]: Started Session c2 of User root. Nov 27 03:13:39 localhost podman[63000]: 2025-11-27 08:13:39.657790849 +0000 UTC m=+0.122138636 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=starting, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, release=1761123044, tcib_managed=true, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3) Nov 27 03:13:39 localhost systemd[1]: session-c1.scope: Deactivated successfully. Nov 27 03:13:39 localhost systemd[1]: session-c2.scope: Deactivated successfully. Nov 27 03:13:39 localhost podman[63000]: 2025-11-27 08:13:39.741806634 +0000 UTC m=+0.206154421 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, architecture=x86_64, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., version=17.1.12, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, distribution-scope=public, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:13:39 localhost podman[63000]: unhealthy Nov 27 03:13:39 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:13:39 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Failed with result 'exit-code'. Nov 27 03:13:39 localhost podman[63170]: 2025-11-27 08:13:39.917949102 +0000 UTC m=+0.056002231 container create 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:35:22Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-libvirt-container, name=rhosp17/openstack-nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, architecture=x86_64) Nov 27 03:13:39 localhost systemd[1]: Started libpod-conmon-6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e.scope. Nov 27 03:13:39 localhost systemd[1]: Started libcrun container. Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d0f444e3451054fa27276f38df962bce4740a4eeaa689d937364bf82d226519f/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d0f444e3451054fa27276f38df962bce4740a4eeaa689d937364bf82d226519f/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d0f444e3451054fa27276f38df962bce4740a4eeaa689d937364bf82d226519f/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d0f444e3451054fa27276f38df962bce4740a4eeaa689d937364bf82d226519f/merged/var/log/swtpm/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:39 localhost podman[63170]: 2025-11-27 08:13:39.976873194 +0000 UTC m=+0.114926313 container init 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:35:22Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, batch=17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true, com.redhat.component=openstack-nova-libvirt-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vendor=Red Hat, Inc., io.buildah.version=1.41.4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, maintainer=OpenStack TripleO Team) Nov 27 03:13:39 localhost podman[63170]: 2025-11-27 08:13:39.983551161 +0000 UTC m=+0.121604280 container start 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, name=rhosp17/openstack-nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, io.buildah.version=1.41.4, distribution-scope=public, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:35:22Z, com.redhat.component=openstack-nova-libvirt-container) Nov 27 03:13:39 localhost podman[63170]: 2025-11-27 08:13:39.891664535 +0000 UTC m=+0.029717754 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:39 localhost podman[63192]: 2025-11-27 08:13:39.99813701 +0000 UTC m=+0.077276097 container create 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-nova-libvirt, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vendor=Red Hat, Inc., config_id=tripleo_step3, com.redhat.component=openstack-nova-libvirt-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, container_name=nova_virtsecretd, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:35:22Z, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4) Nov 27 03:13:40 localhost systemd[1]: Started libpod-conmon-894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c.scope. Nov 27 03:13:40 localhost systemd[1]: Started libcrun container. Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost podman[63192]: 2025-11-27 08:13:39.952438909 +0000 UTC m=+0.031578026 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost podman[63192]: 2025-11-27 08:13:40.05663099 +0000 UTC m=+0.135770077 container init 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, build-date=2025-11-19T00:35:22Z, container_name=nova_virtsecretd, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-libvirt-container, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_step3, version=17.1.12, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt) Nov 27 03:13:40 localhost podman[63192]: 2025-11-27 08:13:40.063018168 +0000 UTC m=+0.142157245 container start 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-libvirt, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, build-date=2025-11-19T00:35:22Z, container_name=nova_virtsecretd, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, io.buildah.version=1.41.4, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public) Nov 27 03:13:40 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtsecretd --cgroupns=host --conmon-pidfile /run/nova_virtsecretd.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtsecretd --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtsecretd.log --network host --pid host --pids-limit 65536 --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:40 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:40 localhost systemd[1]: Started Session c3 of User root. Nov 27 03:13:40 localhost systemd[1]: session-c3.scope: Deactivated successfully. Nov 27 03:13:40 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5-userdata-shm.mount: Deactivated successfully. Nov 27 03:13:40 localhost systemd[1]: var-lib-containers-storage-overlay-7958ccd3e0b0511dc5f6716870f0adbf868f68deb164ea6566f556cebcde5e7f-merged.mount: Deactivated successfully. Nov 27 03:13:40 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-213fb887adc528c7a2b990e624093989af49983e1178b779aad5d9ed39f6be8b-userdata-shm.mount: Deactivated successfully. Nov 27 03:13:40 localhost podman[63336]: 2025-11-27 08:13:40.471951552 +0000 UTC m=+0.091038092 container create bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, io.openshift.expose-services=, version=17.1.12, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.buildah.version=1.41.4, url=https://www.redhat.com, tcib_managed=true, batch=17.1_20251118.1, build-date=2025-11-19T00:35:22Z, container_name=nova_virtnodedevd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-libvirt, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-libvirt-container, vendor=Red Hat, Inc.) Nov 27 03:13:40 localhost systemd[1]: Started libpod-conmon-bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22.scope. Nov 27 03:13:40 localhost podman[63347]: 2025-11-27 08:13:40.514595739 +0000 UTC m=+0.104023248 container create 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, distribution-scope=public, vendor=Red Hat, Inc., version=17.1.12, config_id=tripleo_step3, release=1761123044, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-18T23:44:13Z, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1) Nov 27 03:13:40 localhost podman[63336]: 2025-11-27 08:13:40.427311511 +0000 UTC m=+0.046398101 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:40 localhost systemd[1]: Started libcrun container. Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost systemd[1]: Started libpod-conmon-84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.scope. Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost podman[63336]: 2025-11-27 08:13:40.56067512 +0000 UTC m=+0.179761670 container init bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, url=https://www.redhat.com, architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, com.redhat.component=openstack-nova-libvirt-container, config_id=tripleo_step3, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, build-date=2025-11-19T00:35:22Z, description=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=nova_virtnodedevd, io.buildah.version=1.41.4) Nov 27 03:13:40 localhost podman[63347]: 2025-11-27 08:13:40.463165856 +0000 UTC m=+0.052593405 image pull registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 Nov 27 03:13:40 localhost podman[63336]: 2025-11-27 08:13:40.567586134 +0000 UTC m=+0.186672674 container start bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, managed_by=tripleo_ansible, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, build-date=2025-11-19T00:35:22Z, architecture=x86_64, com.redhat.component=openstack-nova-libvirt-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, container_name=nova_virtnodedevd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, vendor=Red Hat, Inc., config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, name=rhosp17/openstack-nova-libvirt) Nov 27 03:13:40 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtnodedevd --cgroupns=host --conmon-pidfile /run/nova_virtnodedevd.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtnodedevd --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtnodedevd.log --network host --pid host --pids-limit 65536 --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:40 localhost systemd[1]: Started libcrun container. Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/47567cd0527d828c782715e735bfac6bc84c5467ad9fa21661ac755910450a2a/merged/etc/target supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/47567cd0527d828c782715e735bfac6bc84c5467ad9fa21661ac755910450a2a/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:40 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:40 localhost systemd[1]: Started Session c4 of User root. Nov 27 03:13:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:13:40 localhost podman[63347]: 2025-11-27 08:13:40.616983968 +0000 UTC m=+0.206411447 container init 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, batch=17.1_20251118.1, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, io.buildah.version=1.41.4, container_name=iscsid, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, name=rhosp17/openstack-iscsid, vcs-type=git, architecture=x86_64, distribution-scope=public, build-date=2025-11-18T23:44:13Z, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step3) Nov 27 03:13:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:13:40 localhost podman[63347]: 2025-11-27 08:13:40.637309608 +0000 UTC m=+0.226737077 container start 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, architecture=x86_64, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=iscsid, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, distribution-scope=public, io.openshift.expose-services=, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12) Nov 27 03:13:40 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name iscsid --conmon-pidfile /run/iscsid.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=2c87de3317f94758d1bec36af3e86047 --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step3 --label container_name=iscsid --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/iscsid.log --network host --privileged=True --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro --volume /dev:/dev --volume /run:/run --volume /sys:/sys --volume /lib/modules:/lib/modules:ro --volume /var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro --volume /etc/target:/etc/target:z --volume /var/lib/iscsi:/var/lib/iscsi:z registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1 Nov 27 03:13:40 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:40 localhost systemd[1]: Started Session c5 of User root. Nov 27 03:13:40 localhost systemd[1]: session-c4.scope: Deactivated successfully. Nov 27 03:13:40 localhost podman[63393]: 2025-11-27 08:13:40.722385294 +0000 UTC m=+0.076583019 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=starting, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, architecture=x86_64, tcib_managed=true, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, container_name=iscsid, build-date=2025-11-18T23:44:13Z) Nov 27 03:13:40 localhost systemd[1]: session-c5.scope: Deactivated successfully. Nov 27 03:13:40 localhost kernel: Loading iSCSI transport class v2.0-870. Nov 27 03:13:40 localhost podman[63393]: 2025-11-27 08:13:40.75901237 +0000 UTC m=+0.113210115 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, distribution-scope=public, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, container_name=iscsid, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com) Nov 27 03:13:40 localhost podman[63393]: unhealthy Nov 27 03:13:40 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:13:40 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Failed with result 'exit-code'. Nov 27 03:13:41 localhost podman[63516]: 2025-11-27 08:13:41.093398044 +0000 UTC m=+0.088340207 container create c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.buildah.version=1.41.4, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, vcs-type=git, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, release=1761123044, architecture=x86_64, build-date=2025-11-19T00:35:22Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=nova_virtstoraged, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, tcib_managed=true, version=17.1.12, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:13:41 localhost systemd[1]: Started libpod-conmon-c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4.scope. Nov 27 03:13:41 localhost podman[63516]: 2025-11-27 08:13:41.049647608 +0000 UTC m=+0.044589801 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:41 localhost systemd[1]: Started libcrun container. Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost podman[63516]: 2025-11-27 08:13:41.164653762 +0000 UTC m=+0.159595925 container init c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, release=1761123044, architecture=x86_64, build-date=2025-11-19T00:35:22Z, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=nova_virtstoraged, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step3, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-libvirt-container, distribution-scope=public, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, url=https://www.redhat.com, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt) Nov 27 03:13:41 localhost podman[63516]: 2025-11-27 08:13:41.174036195 +0000 UTC m=+0.168978358 container start c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, release=1761123044, architecture=x86_64, io.openshift.expose-services=, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_virtstoraged, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vcs-type=git, config_id=tripleo_step3, com.redhat.component=openstack-nova-libvirt-container, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, io.buildah.version=1.41.4, build-date=2025-11-19T00:35:22Z, vendor=Red Hat, Inc., url=https://www.redhat.com) Nov 27 03:13:41 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtstoraged --cgroupns=host --conmon-pidfile /run/nova_virtstoraged.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtstoraged --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtstoraged.log --network host --pid host --pids-limit 65536 --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:41 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:41 localhost systemd[1]: Started Session c6 of User root. Nov 27 03:13:41 localhost systemd[1]: session-c6.scope: Deactivated successfully. Nov 27 03:13:41 localhost podman[63621]: 2025-11-27 08:13:41.588052502 +0000 UTC m=+0.078324837 container create c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, version=17.1.12, com.redhat.component=openstack-nova-libvirt-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, build-date=2025-11-19T00:35:22Z, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, config_id=tripleo_step3, tcib_managed=true, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., container_name=nova_virtqemud, distribution-scope=public) Nov 27 03:13:41 localhost systemd[1]: Started libpod-conmon-c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb.scope. Nov 27 03:13:41 localhost podman[63621]: 2025-11-27 08:13:41.543357909 +0000 UTC m=+0.033630264 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:41 localhost systemd[1]: Started libcrun container. Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/log/swtpm supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:41 localhost podman[63621]: 2025-11-27 08:13:41.663873327 +0000 UTC m=+0.154145652 container init c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, distribution-scope=public, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, container_name=nova_virtqemud, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, com.redhat.component=openstack-nova-libvirt-container, version=17.1.12, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:13:41 localhost podman[63621]: 2025-11-27 08:13:41.675563935 +0000 UTC m=+0.165836260 container start c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-libvirt, tcib_managed=true, config_id=tripleo_step3, container_name=nova_virtqemud, com.redhat.component=openstack-nova-libvirt-container, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vendor=Red Hat, Inc., io.buildah.version=1.41.4, distribution-scope=public, url=https://www.redhat.com, architecture=x86_64, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:13:41 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtqemud --cgroupns=host --conmon-pidfile /run/nova_virtqemud.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtqemud --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtqemud.log --network host --pid host --pids-limit 65536 --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro --volume /var/log/containers/libvirt/swtpm:/var/log/swtpm:z registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:41 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:41 localhost systemd[1]: Started Session c7 of User root. Nov 27 03:13:41 localhost systemd[1]: session-c7.scope: Deactivated successfully. Nov 27 03:13:42 localhost podman[63728]: 2025-11-27 08:13:42.123907634 +0000 UTC m=+0.084220932 container create 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=nova_virtproxyd, vcs-type=git, release=1761123044, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, com.redhat.component=openstack-nova-libvirt-container, batch=17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, version=17.1.12, name=rhosp17/openstack-nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, io.openshift.expose-services=) Nov 27 03:13:42 localhost systemd[1]: Started libpod-conmon-6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172.scope. Nov 27 03:13:42 localhost systemd[1]: Started libcrun container. Nov 27 03:13:42 localhost podman[63728]: 2025-11-27 08:13:42.077783741 +0000 UTC m=+0.038097069 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/log/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/cache/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/lib/vhost_sockets supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:13:42 localhost podman[63728]: 2025-11-27 08:13:42.185595353 +0000 UTC m=+0.145908651 container init 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, build-date=2025-11-19T00:35:22Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_step3, tcib_managed=true, com.redhat.component=openstack-nova-libvirt-container, container_name=nova_virtproxyd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, batch=17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-libvirt, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 03:13:42 localhost podman[63728]: 2025-11-27 08:13:42.195925403 +0000 UTC m=+0.156238711 container start 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, name=rhosp17/openstack-nova-libvirt, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_virtproxyd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, maintainer=OpenStack TripleO Team, vcs-type=git, build-date=2025-11-19T00:35:22Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-libvirt-container, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step3, tcib_managed=true, release=1761123044) Nov 27 03:13:42 localhost python3[62618]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_virtproxyd --cgroupns=host --conmon-pidfile /run/nova_virtproxyd.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step3 --label container_name=nova_virtproxyd --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_virtproxyd.log --network host --pid host --pids-limit 65536 --privileged=True --security-opt label=level:s0 --security-opt label=type:spc_t --security-opt label=filetype:container_file_t --ulimit nofile=131072 --ulimit nproc=126960 --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/libvirt:/var/log/libvirt:shared,z --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /run:/run --volume /sys/fs/cgroup:/sys/fs/cgroup --volume /sys/fs/selinux:/sys/fs/selinux --volume /etc/selinux/config:/etc/selinux/config:ro --volume /etc/libvirt:/etc/libvirt:shared --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/cache/libvirt:/var/cache/libvirt:shared --volume /var/lib/vhost_sockets:/var/lib/vhost_sockets --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:13:42 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:13:42 localhost systemd[1]: Started Session c8 of User root. Nov 27 03:13:42 localhost systemd[1]: session-c8.scope: Deactivated successfully. Nov 27 03:13:42 localhost python3[63811]: ansible-file Invoked with path=/etc/systemd/system/tripleo_collectd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:43 localhost python3[63827]: ansible-file Invoked with path=/etc/systemd/system/tripleo_iscsid.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:43 localhost python3[63843]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtlogd_wrapper.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:43 localhost python3[63859]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtnodedevd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:43 localhost python3[63875]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtproxyd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:44 localhost python3[63891]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:44 localhost python3[63907]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtsecretd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:44 localhost python3[63923]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_virtstoraged.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:45 localhost python3[63939]: ansible-file Invoked with path=/etc/systemd/system/tripleo_rsyslog.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:45 localhost python3[63955]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_collectd_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:45 localhost python3[63971]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_iscsid_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:45 localhost python3[63987]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtlogd_wrapper_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:46 localhost python3[64003]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtnodedevd_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:46 localhost python3[64019]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtproxyd_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:46 localhost python3[64035]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:47 localhost python3[64051]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtsecretd_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:47 localhost python3[64067]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_virtstoraged_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:47 localhost python3[64083]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_rsyslog_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:13:48 localhost python3[64144]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_collectd.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:13:48 localhost systemd[1]: tmp-crun.eJ54K1.mount: Deactivated successfully. Nov 27 03:13:48 localhost podman[64174]: 2025-11-27 08:13:48.818335926 +0000 UTC m=+0.095654343 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, build-date=2025-11-18T22:49:46Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team) Nov 27 03:13:48 localhost python3[64173]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_iscsid.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:48 localhost podman[64174]: 2025-11-27 08:13:48.996213532 +0000 UTC m=+0.273531959 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.openshift.expose-services=, tcib_managed=true) Nov 27 03:13:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:13:49 localhost python3[64231]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtlogd_wrapper.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:50 localhost python3[64260]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtnodedevd.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:50 localhost python3[64289]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtproxyd.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:51 localhost python3[64318]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtqemud.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:51 localhost python3[64347]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtsecretd.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:52 localhost python3[64376]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_nova_virtstoraged.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:52 localhost systemd[1]: Stopping User Manager for UID 0... Nov 27 03:13:52 localhost systemd[62955]: Activating special unit Exit the Session... Nov 27 03:13:52 localhost systemd[62955]: Stopped target Main User Target. Nov 27 03:13:52 localhost systemd[62955]: Stopped target Basic System. Nov 27 03:13:52 localhost systemd[62955]: Stopped target Paths. Nov 27 03:13:52 localhost systemd[62955]: Stopped target Sockets. Nov 27 03:13:52 localhost systemd[62955]: Stopped target Timers. Nov 27 03:13:52 localhost systemd[62955]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 03:13:52 localhost systemd[62955]: Closed D-Bus User Message Bus Socket. Nov 27 03:13:52 localhost systemd[62955]: Stopped Create User's Volatile Files and Directories. Nov 27 03:13:52 localhost systemd[62955]: Removed slice User Application Slice. Nov 27 03:13:52 localhost systemd[62955]: Reached target Shutdown. Nov 27 03:13:52 localhost systemd[62955]: Finished Exit the Session. Nov 27 03:13:52 localhost systemd[62955]: Reached target Exit the Session. Nov 27 03:13:52 localhost systemd[1]: user@0.service: Deactivated successfully. Nov 27 03:13:52 localhost systemd[1]: Stopped User Manager for UID 0. Nov 27 03:13:52 localhost systemd[1]: Stopping User Runtime Directory /run/user/0... Nov 27 03:13:52 localhost systemd[1]: run-user-0.mount: Deactivated successfully. Nov 27 03:13:52 localhost systemd[1]: user-runtime-dir@0.service: Deactivated successfully. Nov 27 03:13:52 localhost systemd[1]: Stopped User Runtime Directory /run/user/0. Nov 27 03:13:52 localhost systemd[1]: Removed slice User Slice of UID 0. Nov 27 03:13:52 localhost python3[64407]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231227.7437809-100394-238730400225990/source dest=/etc/systemd/system/tripleo_rsyslog.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:13:53 localhost python3[64423]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 03:13:53 localhost systemd[1]: Reloading. Nov 27 03:13:53 localhost systemd-sysv-generator[64451]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:53 localhost systemd-rc-local-generator[64448]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:53 localhost python3[64476]: ansible-systemd Invoked with state=restarted name=tripleo_collectd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:53 localhost systemd[1]: Reloading. Nov 27 03:13:53 localhost systemd-rc-local-generator[64503]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:53 localhost systemd-sysv-generator[64510]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:54 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:54 localhost systemd[1]: Starting dnf makecache... Nov 27 03:13:54 localhost systemd[1]: Starting collectd container... Nov 27 03:13:54 localhost systemd[1]: Started collectd container. Nov 27 03:13:54 localhost dnf[64516]: Updating Subscription Management repositories. Nov 27 03:13:54 localhost python3[64545]: ansible-systemd Invoked with state=restarted name=tripleo_iscsid.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:54 localhost systemd[1]: Reloading. Nov 27 03:13:55 localhost systemd-rc-local-generator[64568]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:55 localhost systemd-sysv-generator[64571]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:55 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:55 localhost systemd[1]: Starting iscsid container... Nov 27 03:13:55 localhost systemd[1]: Started iscsid container. Nov 27 03:13:56 localhost python3[64612]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtlogd_wrapper.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:56 localhost systemd[1]: Reloading. Nov 27 03:13:56 localhost dnf[64516]: Failed determining last makecache time. Nov 27 03:13:56 localhost systemd-sysv-generator[64638]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:56 localhost systemd-rc-local-generator[64635]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:56 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:56 localhost dnf[64516]: Red Hat Enterprise Linux 9 for x86_64 - AppStre 40 kB/s | 4.5 kB 00:00 Nov 27 03:13:56 localhost systemd[1]: Starting nova_virtlogd_wrapper container... Nov 27 03:13:56 localhost systemd[1]: Started nova_virtlogd_wrapper container. Nov 27 03:13:56 localhost dnf[64516]: Red Hat Enterprise Linux 9 for x86_64 - AppStre 48 kB/s | 4.5 kB 00:00 Nov 27 03:13:56 localhost dnf[64516]: Red Hat Enterprise Linux 9 for x86_64 - High Av 44 kB/s | 4.0 kB 00:00 Nov 27 03:13:56 localhost dnf[64516]: Fast Datapath for RHEL 9 x86_64 (RPMs) 26 kB/s | 4.0 kB 00:00 Nov 27 03:13:56 localhost dnf[64516]: Red Hat Enterprise Linux 9 for x86_64 - BaseOS 47 kB/s | 4.1 kB 00:00 Nov 27 03:13:57 localhost python3[64681]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtnodedevd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:57 localhost systemd[1]: Reloading. Nov 27 03:13:57 localhost dnf[64516]: Red Hat OpenStack Platform 17.1 for RHEL 9 x86_ 50 kB/s | 4.0 kB 00:00 Nov 27 03:13:57 localhost systemd-sysv-generator[64714]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:57 localhost systemd-rc-local-generator[64711]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:57 localhost dnf[64516]: Red Hat Enterprise Linux 9 for x86_64 - BaseOS 42 kB/s | 4.1 kB 00:00 Nov 27 03:13:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:57 localhost systemd[1]: Starting nova_virtnodedevd container... Nov 27 03:13:57 localhost dnf[64516]: Metadata cache created. Nov 27 03:13:57 localhost tripleo-start-podman-container[64724]: Creating additional drop-in dependency for "nova_virtnodedevd" (bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22) Nov 27 03:13:57 localhost systemd[1]: Reloading. Nov 27 03:13:57 localhost systemd-sysv-generator[64780]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:57 localhost systemd-rc-local-generator[64777]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:57 localhost systemd[1]: dnf-makecache.service: Deactivated successfully. Nov 27 03:13:57 localhost systemd[1]: Finished dnf makecache. Nov 27 03:13:57 localhost systemd[1]: dnf-makecache.service: Consumed 2.845s CPU time. Nov 27 03:13:57 localhost systemd[1]: Started nova_virtnodedevd container. Nov 27 03:13:58 localhost python3[64807]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtproxyd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:58 localhost systemd[1]: Reloading. Nov 27 03:13:58 localhost systemd-rc-local-generator[64836]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:58 localhost systemd-sysv-generator[64839]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:58 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:58 localhost systemd[1]: Starting nova_virtproxyd container... Nov 27 03:13:59 localhost tripleo-start-podman-container[64847]: Creating additional drop-in dependency for "nova_virtproxyd" (6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172) Nov 27 03:13:59 localhost systemd[1]: Reloading. Nov 27 03:13:59 localhost systemd-sysv-generator[64906]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:13:59 localhost systemd-rc-local-generator[64902]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:13:59 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:13:59 localhost systemd[1]: Started nova_virtproxyd container. Nov 27 03:13:59 localhost python3[64930]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtqemud.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:13:59 localhost systemd[1]: Reloading. Nov 27 03:14:00 localhost systemd-sysv-generator[64958]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:00 localhost systemd-rc-local-generator[64955]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:00 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:00 localhost systemd[1]: Starting nova_virtqemud container... Nov 27 03:14:00 localhost tripleo-start-podman-container[64970]: Creating additional drop-in dependency for "nova_virtqemud" (c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb) Nov 27 03:14:00 localhost systemd[1]: Reloading. Nov 27 03:14:00 localhost systemd-rc-local-generator[65024]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:00 localhost systemd-sysv-generator[65029]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:00 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:00 localhost systemd[1]: Started nova_virtqemud container. Nov 27 03:14:01 localhost python3[65054]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtsecretd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:14:01 localhost systemd[1]: Reloading. Nov 27 03:14:01 localhost systemd-rc-local-generator[65080]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:01 localhost systemd-sysv-generator[65083]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:01 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:01 localhost systemd[1]: Starting nova_virtsecretd container... Nov 27 03:14:01 localhost tripleo-start-podman-container[65093]: Creating additional drop-in dependency for "nova_virtsecretd" (894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c) Nov 27 03:14:01 localhost systemd[1]: Reloading. Nov 27 03:14:01 localhost systemd-sysv-generator[65152]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:01 localhost systemd-rc-local-generator[65149]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:02 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:02 localhost systemd[1]: Started nova_virtsecretd container. Nov 27 03:14:02 localhost python3[65176]: ansible-systemd Invoked with state=restarted name=tripleo_nova_virtstoraged.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:14:03 localhost systemd[1]: Reloading. Nov 27 03:14:03 localhost systemd-rc-local-generator[65206]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:03 localhost systemd-sysv-generator[65209]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:03 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:03 localhost systemd[1]: Starting nova_virtstoraged container... Nov 27 03:14:03 localhost tripleo-start-podman-container[65216]: Creating additional drop-in dependency for "nova_virtstoraged" (c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4) Nov 27 03:14:03 localhost systemd[1]: Reloading. Nov 27 03:14:03 localhost systemd-sysv-generator[65272]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:03 localhost systemd-rc-local-generator[65268]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:03 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:03 localhost systemd[1]: Started nova_virtstoraged container. Nov 27 03:14:04 localhost python3[65299]: ansible-systemd Invoked with state=restarted name=tripleo_rsyslog.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:14:04 localhost systemd[1]: Reloading. Nov 27 03:14:04 localhost systemd-rc-local-generator[65329]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:14:04 localhost systemd-sysv-generator[65333]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:14:04 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:14:04 localhost systemd[1]: Starting rsyslog container... Nov 27 03:14:04 localhost systemd[1]: Started libcrun container. Nov 27 03:14:04 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:04 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:04 localhost podman[65339]: 2025-11-27 08:14:04.953402597 +0000 UTC m=+0.126284002 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, release=1761123044, tcib_managed=true, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=rsyslog, build-date=2025-11-18T22:49:49Z, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.component=openstack-rsyslog-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 rsyslog, io.buildah.version=1.41.4, name=rhosp17/openstack-rsyslog, config_id=tripleo_step3, version=17.1.12) Nov 27 03:14:04 localhost podman[65339]: 2025-11-27 08:14:04.968786749 +0000 UTC m=+0.141668164 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:49Z, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, name=rhosp17/openstack-rsyslog, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step3, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, container_name=rsyslog, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-rsyslog-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, summary=Red Hat OpenStack Platform 17.1 rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:14:04 localhost podman[65339]: rsyslog Nov 27 03:14:04 localhost systemd[1]: Started rsyslog container. Nov 27 03:14:05 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:14:05 localhost podman[65375]: 2025-11-27 08:14:05.130563934 +0000 UTC m=+0.041779142 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, container_name=rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-rsyslog, managed_by=tripleo_ansible, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, url=https://www.redhat.com, distribution-scope=public, build-date=2025-11-18T22:49:49Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.expose-services=, com.redhat.component=openstack-rsyslog-container, config_id=tripleo_step3, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:14:05 localhost podman[65375]: 2025-11-27 08:14:05.155010059 +0000 UTC m=+0.066225227 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, build-date=2025-11-18T22:49:49Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, managed_by=tripleo_ansible, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, container_name=rsyslog, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, version=17.1.12, name=rhosp17/openstack-rsyslog, architecture=x86_64, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, com.redhat.component=openstack-rsyslog-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog) Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:14:05 localhost podman[65389]: 2025-11-27 08:14:05.244249481 +0000 UTC m=+0.056174016 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, container_name=rsyslog, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, release=1761123044, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, description=Red Hat OpenStack Platform 17.1 rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-rsyslog-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, architecture=x86_64, name=rhosp17/openstack-rsyslog, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, distribution-scope=public, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:14:05 localhost podman[65389]: rsyslog Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Scheduled restart job, restart counter is at 1. Nov 27 03:14:05 localhost systemd[1]: Stopped rsyslog container. Nov 27 03:14:05 localhost systemd[1]: Starting rsyslog container... Nov 27 03:14:05 localhost python3[65417]: ansible-file Invoked with path=/var/lib/container-puppet/container-puppet-tasks3.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:14:05 localhost systemd[1]: Started libcrun container. Nov 27 03:14:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:05 localhost podman[65418]: 2025-11-27 08:14:05.52960696 +0000 UTC m=+0.114900372 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, architecture=x86_64, config_id=tripleo_step3, url=https://www.redhat.com, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, release=1761123044, summary=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.buildah.version=1.41.4, com.redhat.component=openstack-rsyslog-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, name=rhosp17/openstack-rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, container_name=rsyslog, build-date=2025-11-18T22:49:49Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, vendor=Red Hat, Inc.) Nov 27 03:14:05 localhost podman[65418]: 2025-11-27 08:14:05.539991112 +0000 UTC m=+0.125284464 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, architecture=x86_64, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, distribution-scope=public, io.buildah.version=1.41.4, vendor=Red Hat, Inc., com.redhat.component=openstack-rsyslog-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T22:49:49Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, container_name=rsyslog, name=rhosp17/openstack-rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, managed_by=tripleo_ansible, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:14:05 localhost podman[65418]: rsyslog Nov 27 03:14:05 localhost systemd[1]: Started rsyslog container. Nov 27 03:14:05 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:14:05 localhost podman[65440]: 2025-11-27 08:14:05.698454584 +0000 UTC m=+0.050132777 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, tcib_managed=true, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, com.redhat.component=openstack-rsyslog-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, config_id=tripleo_step3, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, container_name=rsyslog, name=rhosp17/openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64) Nov 27 03:14:05 localhost podman[65440]: 2025-11-27 08:14:05.72719685 +0000 UTC m=+0.078875003 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, managed_by=tripleo_ansible, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-rsyslog-container, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, name=rhosp17/openstack-rsyslog, url=https://www.redhat.com, architecture=x86_64, vcs-type=git, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, build-date=2025-11-18T22:49:49Z, container_name=rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 rsyslog) Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:14:05 localhost podman[65454]: 2025-11-27 08:14:05.806815632 +0000 UTC m=+0.051643628 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, build-date=2025-11-18T22:49:49Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, name=rhosp17/openstack-rsyslog, architecture=x86_64, com.redhat.component=openstack-rsyslog-container, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, config_id=tripleo_step3, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, release=1761123044, vcs-type=git, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1) Nov 27 03:14:05 localhost podman[65454]: rsyslog Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:05 localhost systemd[1]: tmp-crun.zvhidM.mount: Deactivated successfully. Nov 27 03:14:05 localhost systemd[1]: var-lib-containers-storage-overlay-0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682-merged.mount: Deactivated successfully. Nov 27 03:14:05 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5-userdata-shm.mount: Deactivated successfully. Nov 27 03:14:05 localhost systemd[1]: tripleo_rsyslog.service: Scheduled restart job, restart counter is at 2. Nov 27 03:14:05 localhost systemd[1]: Stopped rsyslog container. Nov 27 03:14:05 localhost systemd[1]: Starting rsyslog container... Nov 27 03:14:06 localhost systemd[1]: tmp-crun.HfYJOG.mount: Deactivated successfully. Nov 27 03:14:06 localhost systemd[1]: Started libcrun container. Nov 27 03:14:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:06 localhost podman[65513]: 2025-11-27 08:14:06.100788343 +0000 UTC m=+0.116763685 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, maintainer=OpenStack TripleO Team, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.openshift.expose-services=, release=1761123044, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 rsyslog, container_name=rsyslog, com.redhat.component=openstack-rsyslog-container, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, name=rhosp17/openstack-rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.buildah.version=1.41.4, config_id=tripleo_step3, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:14:06 localhost podman[65513]: 2025-11-27 08:14:06.10960907 +0000 UTC m=+0.125584402 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, name=rhosp17/openstack-rsyslog, container_name=rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:49Z, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.component=openstack-rsyslog-container, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, architecture=x86_64, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:14:06 localhost podman[65513]: rsyslog Nov 27 03:14:06 localhost systemd[1]: Started rsyslog container. Nov 27 03:14:06 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:14:06 localhost podman[65549]: 2025-11-27 08:14:06.275329436 +0000 UTC m=+0.049101977 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, url=https://www.redhat.com, io.buildah.version=1.41.4, release=1761123044, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, build-date=2025-11-18T22:49:49Z, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-rsyslog-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, managed_by=tripleo_ansible, version=17.1.12, container_name=rsyslog) Nov 27 03:14:06 localhost podman[65549]: 2025-11-27 08:14:06.300670346 +0000 UTC m=+0.074442847 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, architecture=x86_64, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, version=17.1.12, config_id=tripleo_step3, container_name=rsyslog, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.component=openstack-rsyslog-container, io.openshift.expose-services=, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:49Z, url=https://www.redhat.com, name=rhosp17/openstack-rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:14:06 localhost systemd[1]: tripleo_rsyslog.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:14:06 localhost podman[65590]: 2025-11-27 08:14:06.363019835 +0000 UTC m=+0.041427613 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, distribution-scope=public, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=rsyslog, build-date=2025-11-18T22:49:49Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-rsyslog-container, name=rhosp17/openstack-rsyslog, vendor=Red Hat, Inc., version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, url=https://www.redhat.com, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, config_id=tripleo_step3, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:14:06 localhost podman[65590]: rsyslog Nov 27 03:14:06 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:06 localhost systemd[1]: tripleo_rsyslog.service: Scheduled restart job, restart counter is at 3. Nov 27 03:14:06 localhost systemd[1]: Stopped rsyslog container. Nov 27 03:14:06 localhost systemd[1]: Starting rsyslog container... Nov 27 03:14:06 localhost systemd[1]: Started libcrun container. Nov 27 03:14:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:06 localhost podman[65618]: 2025-11-27 08:14:06.769308845 +0000 UTC m=+0.114071499 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, name=rhosp17/openstack-rsyslog, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.component=openstack-rsyslog-container, build-date=2025-11-18T22:49:49Z, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 rsyslog, distribution-scope=public, container_name=rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, url=https://www.redhat.com, version=17.1.12, release=1761123044) Nov 27 03:14:06 localhost podman[65618]: 2025-11-27 08:14:06.778093871 +0000 UTC m=+0.122856515 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, build-date=2025-11-18T22:49:49Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, release=1761123044, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, architecture=x86_64, name=rhosp17/openstack-rsyslog, container_name=rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-rsyslog-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, io.openshift.expose-services=, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog) Nov 27 03:14:06 localhost podman[65618]: rsyslog Nov 27 03:14:06 localhost systemd[1]: Started rsyslog container. Nov 27 03:14:06 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:14:06 localhost podman[65656]: 2025-11-27 08:14:06.924204097 +0000 UTC m=+0.043117369 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, config_id=tripleo_step3, io.buildah.version=1.41.4, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, build-date=2025-11-18T22:49:49Z, name=rhosp17/openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.component=openstack-rsyslog-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 rsyslog, container_name=rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12) Nov 27 03:14:06 localhost python3[65650]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=True puppet_config=/var/lib/container-puppet/container-puppet-tasks3.json short_hostname=np0005537446 step=3 update_config_hash_only=False Nov 27 03:14:06 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5-userdata-shm.mount: Deactivated successfully. Nov 27 03:14:06 localhost systemd[1]: var-lib-containers-storage-overlay-0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682-merged.mount: Deactivated successfully. Nov 27 03:14:06 localhost podman[65656]: 2025-11-27 08:14:06.947753607 +0000 UTC m=+0.066666839 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, tcib_managed=true, config_id=tripleo_step3, distribution-scope=public, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, architecture=x86_64, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, name=rhosp17/openstack-rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, batch=17.1_20251118.1, com.redhat.component=openstack-rsyslog-container, build-date=2025-11-18T22:49:49Z, container_name=rsyslog) Nov 27 03:14:06 localhost systemd[1]: tripleo_rsyslog.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:14:07 localhost podman[65668]: 2025-11-27 08:14:07.040976211 +0000 UTC m=+0.069325915 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, release=1761123044, vcs-type=git, container_name=rsyslog, name=rhosp17/openstack-rsyslog, description=Red Hat OpenStack Platform 17.1 rsyslog, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 rsyslog, version=17.1.12, build-date=2025-11-18T22:49:49Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, tcib_managed=true, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-rsyslog-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}) Nov 27 03:14:07 localhost podman[65668]: rsyslog Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Scheduled restart job, restart counter is at 4. Nov 27 03:14:07 localhost systemd[1]: Stopped rsyslog container. Nov 27 03:14:07 localhost systemd[1]: Starting rsyslog container... Nov 27 03:14:07 localhost systemd[1]: Started libcrun container. Nov 27 03:14:07 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/log/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:07 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0ff5e222838f05b2cf66c1e9468b5029e3e46fcc9a4de06559cf54c6751dd682/merged/var/lib/rsyslog supports timestamps until 2038 (0x7fffffff) Nov 27 03:14:07 localhost podman[65696]: 2025-11-27 08:14:07.468610769 +0000 UTC m=+0.123652438 container init 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, com.redhat.component=openstack-rsyslog-container, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., release=1761123044, url=https://www.redhat.com, name=rhosp17/openstack-rsyslog, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, vcs-type=git, build-date=2025-11-18T22:49:49Z, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, container_name=rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, version=17.1.12) Nov 27 03:14:07 localhost podman[65696]: 2025-11-27 08:14:07.476900131 +0000 UTC m=+0.131941810 container start 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, name=rhosp17/openstack-rsyslog, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, version=17.1.12, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-rsyslog-container, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, build-date=2025-11-18T22:49:49Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 rsyslog, container_name=rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, release=1761123044, summary=Red Hat OpenStack Platform 17.1 rsyslog, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog) Nov 27 03:14:07 localhost podman[65696]: rsyslog Nov 27 03:14:07 localhost systemd[1]: Started rsyslog container. Nov 27 03:14:07 localhost python3[65695]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:14:07 localhost systemd[1]: libpod-3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5.scope: Deactivated successfully. Nov 27 03:14:07 localhost podman[65716]: 2025-11-27 08:14:07.659768647 +0000 UTC m=+0.063694216 container died 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, batch=17.1_20251118.1, version=17.1.12, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, managed_by=tripleo_ansible, vcs-type=git, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 rsyslog, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-rsyslog-container, config_id=tripleo_step3, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, summary=Red Hat OpenStack Platform 17.1 rsyslog, url=https://www.redhat.com, name=rhosp17/openstack-rsyslog, build-date=2025-11-18T22:49:49Z) Nov 27 03:14:07 localhost podman[65716]: 2025-11-27 08:14:07.686359583 +0000 UTC m=+0.090285102 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 rsyslog, com.redhat.component=openstack-rsyslog-container, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-rsyslog, release=1761123044, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, batch=17.1_20251118.1, architecture=x86_64, maintainer=OpenStack TripleO Team, container_name=rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, config_id=tripleo_step3, io.openshift.expose-services=, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:49Z, io.buildah.version=1.41.4, version=17.1.12, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog) Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:14:07 localhost podman[65728]: 2025-11-27 08:14:07.759946316 +0000 UTC m=+0.043513530 container cleanup 3aaeb01ad324275c8fbbca5a3bf643153dacc35d1020d1f9b16f910963b57ca5 (image=registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1, name=rsyslog, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 rsyslog, io.buildah.version=1.41.4, name=rhosp17/openstack-rsyslog, com.redhat.component=openstack-rsyslog-container, description=Red Hat OpenStack Platform 17.1 rsyslog, maintainer=OpenStack TripleO Team, container_name=rsyslog, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fd416cb72dcd90fffe349badf5624773'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-rsyslog:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/rsyslog.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/rsyslog:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:ro', '/var/log/containers/rsyslog:/var/log/rsyslog:rw,z', '/var/log:/var/log/host:ro', '/var/lib/rsyslog.container:/var/lib/rsyslog:rw,z']}, summary=Red Hat OpenStack Platform 17.1 rsyslog, io.k8s.display-name=Red Hat OpenStack Platform 17.1 rsyslog, vendor=Red Hat, Inc., tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:49Z, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-rsyslog, batch=17.1_20251118.1) Nov 27 03:14:07 localhost podman[65728]: rsyslog Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Scheduled restart job, restart counter is at 5. Nov 27 03:14:07 localhost systemd[1]: Stopped rsyslog container. Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Start request repeated too quickly. Nov 27 03:14:07 localhost systemd[1]: tripleo_rsyslog.service: Failed with result 'exit-code'. Nov 27 03:14:07 localhost systemd[1]: Failed to start rsyslog container. Nov 27 03:14:07 localhost python3[65754]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_3 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:14:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:14:09 localhost systemd[1]: tmp-crun.qq5Kxi.mount: Deactivated successfully. Nov 27 03:14:10 localhost podman[65755]: 2025-11-27 08:14:10.00272393 +0000 UTC m=+0.102811373 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=starting, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vcs-type=git, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, container_name=collectd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, config_id=tripleo_step3, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:14:10 localhost podman[65755]: 2025-11-27 08:14:10.01411258 +0000 UTC m=+0.114200043 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, build-date=2025-11-18T22:51:28Z, architecture=x86_64, config_id=tripleo_step3, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.expose-services=, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:14:10 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:14:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:14:10 localhost systemd[1]: tmp-crun.Vx0FIm.mount: Deactivated successfully. Nov 27 03:14:10 localhost podman[65775]: 2025-11-27 08:14:10.972054765 +0000 UTC m=+0.068152442 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=starting, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, container_name=iscsid, vendor=Red Hat, Inc., architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z) Nov 27 03:14:11 localhost podman[65775]: 2025-11-27 08:14:11.006830599 +0000 UTC m=+0.102928286 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, container_name=iscsid, vendor=Red Hat, Inc., io.openshift.expose-services=) Nov 27 03:14:11 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:14:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:14:19 localhost systemd[1]: tmp-crun.qMARPz.mount: Deactivated successfully. Nov 27 03:14:19 localhost podman[65794]: 2025-11-27 08:14:19.987770342 +0000 UTC m=+0.089758127 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, container_name=metrics_qdr, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, release=1761123044, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}) Nov 27 03:14:20 localhost podman[65794]: 2025-11-27 08:14:20.171993977 +0000 UTC m=+0.273981762 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, vcs-type=git, url=https://www.redhat.com, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 03:14:20 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:14:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:14:40 localhost systemd[1]: tmp-crun.lR5Nss.mount: Deactivated successfully. Nov 27 03:14:40 localhost podman[65900]: 2025-11-27 08:14:40.972007149 +0000 UTC m=+0.073548946 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, release=1761123044, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, container_name=collectd, config_id=tripleo_step3, version=17.1.12, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git) Nov 27 03:14:40 localhost podman[65900]: 2025-11-27 08:14:40.982154491 +0000 UTC m=+0.083696278 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, release=1761123044, name=rhosp17/openstack-collectd, tcib_managed=true, vendor=Red Hat, Inc., container_name=collectd, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container) Nov 27 03:14:40 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:14:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:14:41 localhost systemd[1]: tmp-crun.SZlAeG.mount: Deactivated successfully. Nov 27 03:14:41 localhost podman[65921]: 2025-11-27 08:14:41.982593456 +0000 UTC m=+0.081548722 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, release=1761123044, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, vcs-type=git, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:14:41 localhost podman[65921]: 2025-11-27 08:14:41.992037799 +0000 UTC m=+0.090993045 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, url=https://www.redhat.com, version=17.1.12, io.buildah.version=1.41.4, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, distribution-scope=public, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, container_name=iscsid, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:14:42 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:14:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:14:50 localhost podman[65941]: 2025-11-27 08:14:50.984874698 +0000 UTC m=+0.080570625 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, url=https://www.redhat.com, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:14:51 localhost podman[65941]: 2025-11-27 08:14:51.200991331 +0000 UTC m=+0.296687228 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, architecture=x86_64, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12) Nov 27 03:14:51 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:15:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:15:11 localhost podman[65969]: 2025-11-27 08:15:11.985086856 +0000 UTC m=+0.078972681 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, distribution-scope=public, version=17.1.12, com.redhat.component=openstack-collectd-container, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git) Nov 27 03:15:11 localhost podman[65969]: 2025-11-27 08:15:11.99343377 +0000 UTC m=+0.087319565 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.41.4, batch=17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, architecture=x86_64, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:15:12 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:15:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:15:12 localhost systemd[1]: tmp-crun.8TA4Ic.mount: Deactivated successfully. Nov 27 03:15:12 localhost podman[65988]: 2025-11-27 08:15:12.106808396 +0000 UTC m=+0.077270587 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_id=tripleo_step3, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, distribution-scope=public, container_name=iscsid, io.buildah.version=1.41.4) Nov 27 03:15:12 localhost podman[65988]: 2025-11-27 08:15:12.140440749 +0000 UTC m=+0.110902940 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-type=git, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=iscsid, io.buildah.version=1.41.4, distribution-scope=public, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, tcib_managed=true, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:15:12 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:15:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:15:21 localhost podman[66022]: 2025-11-27 08:15:21.881601222 +0000 UTC m=+0.135419297 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, tcib_managed=true, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, version=17.1.12, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, distribution-scope=public, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, vcs-type=git, io.buildah.version=1.41.4) Nov 27 03:15:22 localhost podman[66022]: 2025-11-27 08:15:22.110277433 +0000 UTC m=+0.364095568 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, io.openshift.expose-services=, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., release=1761123044, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, batch=17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:15:22 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:15:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:15:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:15:43 localhost systemd[1]: tmp-crun.w9vULN.mount: Deactivated successfully. Nov 27 03:15:43 localhost podman[66117]: 2025-11-27 08:15:43.009620993 +0000 UTC m=+0.098933601 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step3, version=17.1.12, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:15:43 localhost podman[66117]: 2025-11-27 08:15:43.043180748 +0000 UTC m=+0.132493346 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, container_name=iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., version=17.1.12, distribution-scope=public, architecture=x86_64, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, tcib_managed=true, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:15:43 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:15:43 localhost podman[66116]: 2025-11-27 08:15:43.04956538 +0000 UTC m=+0.140778360 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, name=rhosp17/openstack-collectd, tcib_managed=true, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:15:43 localhost podman[66116]: 2025-11-27 08:15:43.129397885 +0000 UTC m=+0.220610865 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, version=17.1.12, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, io.openshift.expose-services=, url=https://www.redhat.com) Nov 27 03:15:43 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:15:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:15:52 localhost systemd[1]: tmp-crun.q5PI9Y.mount: Deactivated successfully. Nov 27 03:15:52 localhost podman[66156]: 2025-11-27 08:15:52.992298666 +0000 UTC m=+0.086790103 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vendor=Red Hat, Inc., batch=17.1_20251118.1, vcs-type=git, tcib_managed=true) Nov 27 03:15:53 localhost podman[66156]: 2025-11-27 08:15:53.185942861 +0000 UTC m=+0.280434288 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, release=1761123044, name=rhosp17/openstack-qdrouterd, managed_by=tripleo_ansible, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:15:53 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:16:10 localhost sshd[66186]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:16:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:16:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:16:13 localhost systemd[1]: tmp-crun.Und8P7.mount: Deactivated successfully. Nov 27 03:16:13 localhost podman[66188]: 2025-11-27 08:16:13.99010709 +0000 UTC m=+0.082520197 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.expose-services=, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, container_name=collectd, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, batch=17.1_20251118.1, vcs-type=git, distribution-scope=public) Nov 27 03:16:14 localhost podman[66188]: 2025-11-27 08:16:14.003793519 +0000 UTC m=+0.096206656 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, architecture=x86_64, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, distribution-scope=public, config_id=tripleo_step3, version=17.1.12, vendor=Red Hat, Inc., batch=17.1_20251118.1, vcs-type=git, release=1761123044, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container) Nov 27 03:16:14 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:16:14 localhost systemd[1]: tmp-crun.Ec6fn4.mount: Deactivated successfully. Nov 27 03:16:14 localhost podman[66189]: 2025-11-27 08:16:14.046158712 +0000 UTC m=+0.136894515 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, container_name=iscsid, batch=17.1_20251118.1, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, release=1761123044, io.openshift.expose-services=, vcs-type=git, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:16:14 localhost podman[66189]: 2025-11-27 08:16:14.058903657 +0000 UTC m=+0.149639420 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid) Nov 27 03:16:14 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:16:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:16:23 localhost systemd[1]: tmp-crun.Z3Hj7p.mount: Deactivated successfully. Nov 27 03:16:23 localhost podman[66240]: 2025-11-27 08:16:23.528845394 +0000 UTC m=+0.083249908 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, managed_by=tripleo_ansible, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, version=17.1.12, tcib_managed=true, distribution-scope=public, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:16:23 localhost podman[66240]: 2025-11-27 08:16:23.758502911 +0000 UTC m=+0.312907345 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, version=17.1.12, batch=17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, url=https://www.redhat.com, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 03:16:23 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:16:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:16:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:16:44 localhost systemd[1]: tmp-crun.PTGJws.mount: Deactivated successfully. Nov 27 03:16:45 localhost podman[66333]: 2025-11-27 08:16:45.00743685 +0000 UTC m=+0.097034019 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-type=git, maintainer=OpenStack TripleO Team, container_name=collectd, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, release=1761123044, version=17.1.12) Nov 27 03:16:45 localhost podman[66333]: 2025-11-27 08:16:45.044424749 +0000 UTC m=+0.134021908 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, architecture=x86_64, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, container_name=collectd, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, tcib_managed=true, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, distribution-scope=public, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, version=17.1.12) Nov 27 03:16:45 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:16:45 localhost podman[66334]: 2025-11-27 08:16:45.087993804 +0000 UTC m=+0.176694899 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, distribution-scope=public, release=1761123044, config_id=tripleo_step3, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:16:45 localhost podman[66334]: 2025-11-27 08:16:45.099843474 +0000 UTC m=+0.188544649 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, io.openshift.expose-services=, tcib_managed=true, release=1761123044) Nov 27 03:16:45 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:16:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:16:53 localhost podman[66370]: 2025-11-27 08:16:53.993333227 +0000 UTC m=+0.092235140 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, container_name=metrics_qdr, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, config_id=tripleo_step1, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., distribution-scope=public, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:16:54 localhost podman[66370]: 2025-11-27 08:16:54.208084772 +0000 UTC m=+0.306986665 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, release=1761123044, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, config_id=tripleo_step1, managed_by=tripleo_ansible, container_name=metrics_qdr, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:16:54 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:17:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:17:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:17:15 localhost podman[66399]: 2025-11-27 08:17:15.987011203 +0000 UTC m=+0.081947952 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, architecture=x86_64, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044) Nov 27 03:17:15 localhost podman[66399]: 2025-11-27 08:17:15.996581061 +0000 UTC m=+0.091517790 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., release=1761123044, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:17:16 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:17:16 localhost systemd[1]: tmp-crun.tdfoaR.mount: Deactivated successfully. Nov 27 03:17:16 localhost podman[66400]: 2025-11-27 08:17:16.089832718 +0000 UTC m=+0.181359695 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.component=openstack-iscsid-container, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, container_name=iscsid, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:17:16 localhost podman[66400]: 2025-11-27 08:17:16.126869847 +0000 UTC m=+0.218396854 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, distribution-scope=public, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., container_name=iscsid, batch=17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:17:16 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:17:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:17:24 localhost podman[66440]: 2025-11-27 08:17:24.974045598 +0000 UTC m=+0.074282025 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, container_name=metrics_qdr, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, version=17.1.12, release=1761123044, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true) Nov 27 03:17:25 localhost podman[66440]: 2025-11-27 08:17:25.159892783 +0000 UTC m=+0.260129210 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, distribution-scope=public, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, tcib_managed=true, config_id=tripleo_step1) Nov 27 03:17:25 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:17:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:17:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:17:47 localhost systemd[1]: tmp-crun.eX1W7n.mount: Deactivated successfully. Nov 27 03:17:47 localhost podman[66596]: 2025-11-27 08:17:47.008820715 +0000 UTC m=+0.097349988 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, release=1761123044, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, config_id=tripleo_step3) Nov 27 03:17:47 localhost podman[66596]: 2025-11-27 08:17:47.022133894 +0000 UTC m=+0.110663197 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, distribution-scope=public, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, com.redhat.component=openstack-iscsid-container, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, io.buildah.version=1.41.4, tcib_managed=true, name=rhosp17/openstack-iscsid, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:17:47 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:17:47 localhost podman[66595]: 2025-11-27 08:17:47.098354971 +0000 UTC m=+0.190298946 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, build-date=2025-11-18T22:51:28Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 03:17:47 localhost podman[66595]: 2025-11-27 08:17:47.135861703 +0000 UTC m=+0.227805588 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, container_name=collectd, vcs-type=git, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, distribution-scope=public, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:17:47 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 1800.1 total, 600.0 interval#012Cumulative writes: 4601 writes, 21K keys, 4601 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4601 writes, 459 syncs, 10.02 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 312 writes, 697 keys, 312 commit groups, 1.0 writes per commit group, ingest: 0.61 MB, 0.00 MB/s#012Interval WAL: 312 writes, 156 syncs, 2.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:17:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:17:55 localhost podman[66634]: 2025-11-27 08:17:55.981152286 +0000 UTC m=+0.081949583 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, tcib_managed=true, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, release=1761123044, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, maintainer=OpenStack TripleO Team, architecture=x86_64) Nov 27 03:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 1800.2 total, 600.0 interval#012Cumulative writes: 4903 writes, 21K keys, 4903 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4903 writes, 543 syncs, 9.03 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 371 writes, 764 keys, 371 commit groups, 1.0 writes per commit group, ingest: 0.59 MB, 0.00 MB/s#012Interval WAL: 371 writes, 185 syncs, 2.01 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:17:56 localhost podman[66634]: 2025-11-27 08:17:56.13474791 +0000 UTC m=+0.235545237 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_id=tripleo_step1, release=1761123044, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, tcib_managed=true, vcs-type=git, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4) Nov 27 03:17:56 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:18:14 localhost python3[66711]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/config_step.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:14 localhost python3[66756]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/config_step.json force=True mode=0600 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231494.0732403-107776-123959030128371/source _original_basename=tmpd4bknnj1 follow=False checksum=ee48fb03297eb703b1954c8852d0f67fab51dac1 backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:16 localhost python3[66818]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/recover_tripleo_nova_virtqemud.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:16 localhost python3[66861]: ansible-ansible.legacy.copy Invoked with dest=/usr/libexec/recover_tripleo_nova_virtqemud.sh mode=0755 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231495.9911337-107874-152572748160465/source _original_basename=tmpd09cyan4 follow=False checksum=922b8aa8342176110bffc2e39abdccc2b39e53a9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:18:17 localhost podman[66924]: 2025-11-27 08:18:17.217481508 +0000 UTC m=+0.080385551 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, url=https://www.redhat.com, distribution-scope=public, architecture=x86_64, container_name=iscsid, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, managed_by=tripleo_ansible, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:18:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:18:17 localhost podman[66924]: 2025-11-27 08:18:17.224046566 +0000 UTC m=+0.086950549 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, container_name=iscsid, io.openshift.expose-services=, url=https://www.redhat.com, vcs-type=git, distribution-scope=public, tcib_managed=true, managed_by=tripleo_ansible, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, release=1761123044, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12) Nov 27 03:18:17 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:18:17 localhost python3[66923]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud_recover.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:17 localhost podman[66943]: 2025-11-27 08:18:17.317849666 +0000 UTC m=+0.087408619 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, version=17.1.12, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, tcib_managed=true, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, container_name=collectd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, architecture=x86_64, io.buildah.version=1.41.4) Nov 27 03:18:17 localhost podman[66943]: 2025-11-27 08:18:17.325795781 +0000 UTC m=+0.095354714 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vcs-type=git, batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:18:17 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:18:17 localhost python3[67005]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/tripleo_nova_virtqemud_recover.service mode=0644 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231496.9581878-107928-211875716969086/source _original_basename=tmp476lzxmm follow=False checksum=92f73544b703afc85885fa63ab07bdf8f8671554 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:18 localhost python3[67067]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud_recover.timer follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:18 localhost python3[67110]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/tripleo_nova_virtqemud_recover.timer mode=0644 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231497.9397395-107984-732679639021/source _original_basename=tmptuvo4059 follow=False checksum=c6e5f76a53c0d6ccaf46c4b48d813dc2891ad8e9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:19 localhost python3[67140]: ansible-systemd Invoked with daemon_reload=True enabled=True name=tripleo_nova_virtqemud_recover.service daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 03:18:19 localhost systemd[1]: Reloading. Nov 27 03:18:19 localhost systemd-sysv-generator[67169]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:19 localhost systemd-rc-local-generator[67164]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:19 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:20 localhost systemd[1]: Reloading. Nov 27 03:18:20 localhost systemd-sysv-generator[67205]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:20 localhost systemd-rc-local-generator[67201]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:20 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:20 localhost python3[67229]: ansible-systemd Invoked with daemon_reload=True enabled=True name=tripleo_nova_virtqemud_recover.timer state=restarted daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:18:21 localhost systemd[1]: Reloading. Nov 27 03:18:21 localhost systemd-rc-local-generator[67251]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:21 localhost systemd-sysv-generator[67255]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:21 localhost systemd[1]: Reloading. Nov 27 03:18:21 localhost systemd-rc-local-generator[67291]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:21 localhost systemd-sysv-generator[67296]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:21 localhost systemd[1]: Started Check and recover tripleo_nova_virtqemud every 10m. Nov 27 03:18:22 localhost python3[67320]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl enable --now tripleo_nova_virtqemud_recover.timer _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 03:18:22 localhost systemd[1]: Reloading. Nov 27 03:18:22 localhost systemd-rc-local-generator[67346]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:22 localhost systemd-sysv-generator[67351]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:22 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:22 localhost python3[67405]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/tripleo_nova_libvirt.target follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:23 localhost python3[67448]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/tripleo_nova_libvirt.target group=root mode=0644 owner=root src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231502.4877126-108153-62570087773510/source _original_basename=tmpp77bzyem follow=False checksum=c064b4a8e7d3d1d7c62d1f80a09e350659996afd backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:23 localhost python3[67478]: ansible-systemd Invoked with daemon_reload=True enabled=True name=tripleo_nova_libvirt.target state=restarted daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:18:23 localhost systemd[1]: Reloading. Nov 27 03:18:23 localhost systemd-sysv-generator[67504]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:23 localhost systemd-rc-local-generator[67499]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:23 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:24 localhost systemd[1]: Reached target tripleo_nova_libvirt.target. Nov 27 03:18:24 localhost python3[67533]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_4 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:18:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:18:26 localhost podman[67706]: 2025-11-27 08:18:26.362180359 +0000 UTC m=+0.087334709 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, architecture=x86_64, distribution-scope=public, managed_by=tripleo_ansible, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, container_name=metrics_qdr, tcib_managed=true) Nov 27 03:18:26 localhost ansible-async_wrapper.py[67705]: Invoked with 563485811991 3600 /home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231505.8570256-108257-79085086938450/AnsiballZ_command.py _ Nov 27 03:18:26 localhost ansible-async_wrapper.py[67737]: Starting module and watcher Nov 27 03:18:26 localhost ansible-async_wrapper.py[67737]: Start watching 67738 (3600) Nov 27 03:18:26 localhost ansible-async_wrapper.py[67738]: Start module (67738) Nov 27 03:18:26 localhost ansible-async_wrapper.py[67705]: Return async_wrapper task started. Nov 27 03:18:26 localhost podman[67706]: 2025-11-27 08:18:26.542797202 +0000 UTC m=+0.267951562 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, release=1761123044, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, managed_by=tripleo_ansible, batch=17.1_20251118.1, url=https://www.redhat.com) Nov 27 03:18:26 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:18:26 localhost python3[67758]: ansible-ansible.legacy.async_status Invoked with jid=563485811991.67705 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:18:29 localhost puppet-user[67756]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:18:29 localhost puppet-user[67756]: (file: /etc/puppet/hiera.yaml) Nov 27 03:18:29 localhost puppet-user[67756]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:18:29 localhost puppet-user[67756]: (file & line not available) Nov 27 03:18:29 localhost puppet-user[67756]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:18:29 localhost puppet-user[67756]: (file & line not available) Nov 27 03:18:29 localhost puppet-user[67756]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/database/mysql/client.pp, line: 89, column: 8) Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use match expressions with Stdlib::Compat::String instead. They are described at https://docs.puppet.com/puppet/latest/reference/lang_data_type.html#match-expressions. at ["/etc/puppet/modules/snmp/manifests/params.pp", 310]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:18:30 localhost puppet-user[67756]: with Stdlib::Compat::Bool. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 358]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:18:30 localhost puppet-user[67756]: with Stdlib::Compat::Array. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 367]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:18:30 localhost puppet-user[67756]: with Stdlib::Compat::String. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 382]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:18:30 localhost puppet-user[67756]: with Stdlib::Compat::Numeric. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 388]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:18:30 localhost puppet-user[67756]: with Pattern[]. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 393]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:18:30 localhost puppet-user[67756]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:18:30 localhost puppet-user[67756]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/packages.pp, line: 39, column: 69) Nov 27 03:18:30 localhost puppet-user[67756]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.25 seconds Nov 27 03:18:31 localhost ansible-async_wrapper.py[67737]: 67738 still running (3600) Nov 27 03:18:36 localhost podman[68128]: Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.113287302 +0000 UTC m=+0.087029329 container create 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, architecture=x86_64, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, ceph=True, name=rhceph, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, GIT_BRANCH=main, maintainer=Guillaume Abrioux , vcs-type=git, version=7, GIT_CLEAN=True, io.openshift.tags=rhceph ceph) Nov 27 03:18:36 localhost systemd[1]: Started libpod-conmon-7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee.scope. Nov 27 03:18:36 localhost systemd[1]: Started libcrun container. Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.080455577 +0000 UTC m=+0.054197634 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.18804832 +0000 UTC m=+0.161790337 container init 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, GIT_BRANCH=main, name=rhceph, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , vcs-type=git, version=7, ceph=True, io.buildah.version=1.33.12, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, RELEASE=main, io.openshift.tags=rhceph ceph, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True) Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.203672852 +0000 UTC m=+0.177414879 container start 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, io.buildah.version=1.33.12, vcs-type=git, GIT_CLEAN=True, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, RELEASE=main, GIT_BRANCH=main, architecture=x86_64, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7) Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.203953209 +0000 UTC m=+0.177695236 container attach 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, io.openshift.expose-services=, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, ceph=True, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_CLEAN=True, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, distribution-scope=public, release=553, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_BRANCH=main, RELEASE=main) Nov 27 03:18:36 localhost zealous_cannon[68143]: 167 167 Nov 27 03:18:36 localhost systemd[1]: libpod-7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee.scope: Deactivated successfully. Nov 27 03:18:36 localhost podman[68128]: 2025-11-27 08:18:36.20845149 +0000 UTC m=+0.182193567 container died 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, version=7, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , vcs-type=git, release=553, distribution-scope=public, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55) Nov 27 03:18:36 localhost podman[68148]: 2025-11-27 08:18:36.284329287 +0000 UTC m=+0.069351451 container remove 7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zealous_cannon, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, architecture=x86_64, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, name=rhceph, version=7, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, distribution-scope=public, GIT_BRANCH=main, RELEASE=main, vendor=Red Hat, Inc., release=553, ceph=True, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 03:18:36 localhost systemd[1]: libpod-conmon-7ccd3e4df2a97a7f41123063a2b376aeaa482e0b7b1f75edb02c5eedbbb376ee.scope: Deactivated successfully. Nov 27 03:18:36 localhost ansible-async_wrapper.py[67737]: 67738 still running (3595) Nov 27 03:18:36 localhost podman[68170]: Nov 27 03:18:36 localhost podman[68170]: 2025-11-27 08:18:36.502261488 +0000 UTC m=+0.076297529 container create 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, version=7, io.buildah.version=1.33.12, io.openshift.expose-services=, distribution-scope=public, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., name=rhceph, maintainer=Guillaume Abrioux , GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main) Nov 27 03:18:36 localhost systemd[1]: Started libpod-conmon-96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5.scope. Nov 27 03:18:36 localhost systemd[1]: Started libcrun container. Nov 27 03:18:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6c9441dafc917f7c408a1453ff9a59e021e1fb773b3c60991d607953b9c46cad/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 03:18:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6c9441dafc917f7c408a1453ff9a59e021e1fb773b3c60991d607953b9c46cad/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:18:36 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6c9441dafc917f7c408a1453ff9a59e021e1fb773b3c60991d607953b9c46cad/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 03:18:36 localhost podman[68170]: 2025-11-27 08:18:36.471084507 +0000 UTC m=+0.045120578 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 03:18:36 localhost podman[68170]: 2025-11-27 08:18:36.572546145 +0000 UTC m=+0.146582196 container init 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, vcs-type=git, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., GIT_BRANCH=main, ceph=True, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, io.openshift.expose-services=, GIT_CLEAN=True, name=rhceph, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 03:18:36 localhost podman[68170]: 2025-11-27 08:18:36.581499616 +0000 UTC m=+0.155535677 container start 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, version=7, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., GIT_BRANCH=main, name=rhceph, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , distribution-scope=public, GIT_CLEAN=True, RELEASE=main, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, release=553, ceph=True) Nov 27 03:18:36 localhost podman[68170]: 2025-11-27 08:18:36.581830636 +0000 UTC m=+0.155866717 container attach 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , release=553, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., version=7, RELEASE=main, io.buildah.version=1.33.12, io.openshift.expose-services=, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, distribution-scope=public, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 03:18:37 localhost python3[68211]: ansible-ansible.legacy.async_status Invoked with jid=563485811991.67705 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:18:37 localhost systemd[1]: tmp-crun.4MMhPw.mount: Deactivated successfully. Nov 27 03:18:37 localhost systemd[1]: var-lib-containers-storage-overlay-eccbd616d00066db957da8d5bb98f89b1de854e37367be247acb79ddfb3c78fe-merged.mount: Deactivated successfully. Nov 27 03:18:37 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 03:18:37 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 03:18:37 localhost systemd[1]: Reloading. Nov 27 03:18:37 localhost systemd-rc-local-generator[69161]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:37 localhost systemd-sysv-generator[69164]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:37 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:37 localhost amazing_taussig[68185]: [ Nov 27 03:18:37 localhost amazing_taussig[68185]: { Nov 27 03:18:37 localhost amazing_taussig[68185]: "available": false, Nov 27 03:18:37 localhost amazing_taussig[68185]: "ceph_device": false, Nov 27 03:18:37 localhost amazing_taussig[68185]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 03:18:37 localhost amazing_taussig[68185]: "lsm_data": {}, Nov 27 03:18:37 localhost amazing_taussig[68185]: "lvs": [], Nov 27 03:18:37 localhost amazing_taussig[68185]: "path": "/dev/sr0", Nov 27 03:18:37 localhost amazing_taussig[68185]: "rejected_reasons": [ Nov 27 03:18:37 localhost amazing_taussig[68185]: "Has a FileSystem", Nov 27 03:18:37 localhost amazing_taussig[68185]: "Insufficient space (<5GB)" Nov 27 03:18:37 localhost amazing_taussig[68185]: ], Nov 27 03:18:37 localhost amazing_taussig[68185]: "sys_api": { Nov 27 03:18:37 localhost amazing_taussig[68185]: "actuators": null, Nov 27 03:18:37 localhost amazing_taussig[68185]: "device_nodes": "sr0", Nov 27 03:18:37 localhost amazing_taussig[68185]: "human_readable_size": "482.00 KB", Nov 27 03:18:37 localhost amazing_taussig[68185]: "id_bus": "ata", Nov 27 03:18:37 localhost amazing_taussig[68185]: "model": "QEMU DVD-ROM", Nov 27 03:18:37 localhost amazing_taussig[68185]: "nr_requests": "2", Nov 27 03:18:37 localhost amazing_taussig[68185]: "partitions": {}, Nov 27 03:18:37 localhost amazing_taussig[68185]: "path": "/dev/sr0", Nov 27 03:18:37 localhost amazing_taussig[68185]: "removable": "1", Nov 27 03:18:37 localhost amazing_taussig[68185]: "rev": "2.5+", Nov 27 03:18:37 localhost amazing_taussig[68185]: "ro": "0", Nov 27 03:18:37 localhost amazing_taussig[68185]: "rotational": "1", Nov 27 03:18:37 localhost amazing_taussig[68185]: "sas_address": "", Nov 27 03:18:37 localhost amazing_taussig[68185]: "sas_device_handle": "", Nov 27 03:18:37 localhost amazing_taussig[68185]: "scheduler_mode": "mq-deadline", Nov 27 03:18:37 localhost amazing_taussig[68185]: "sectors": 0, Nov 27 03:18:37 localhost amazing_taussig[68185]: "sectorsize": "2048", Nov 27 03:18:37 localhost amazing_taussig[68185]: "size": 493568.0, Nov 27 03:18:37 localhost amazing_taussig[68185]: "support_discard": "0", Nov 27 03:18:37 localhost amazing_taussig[68185]: "type": "disk", Nov 27 03:18:37 localhost amazing_taussig[68185]: "vendor": "QEMU" Nov 27 03:18:37 localhost amazing_taussig[68185]: } Nov 27 03:18:37 localhost amazing_taussig[68185]: } Nov 27 03:18:37 localhost amazing_taussig[68185]: ] Nov 27 03:18:37 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 03:18:37 localhost systemd[1]: libpod-96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5.scope: Deactivated successfully. Nov 27 03:18:37 localhost podman[68170]: 2025-11-27 08:18:37.605441597 +0000 UTC m=+1.179477608 container died 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, io.buildah.version=1.33.12, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, version=7, release=553, io.openshift.tags=rhceph ceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, CEPH_POINT_RELEASE=, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55) Nov 27 03:18:37 localhost systemd[1]: var-lib-containers-storage-overlay-6c9441dafc917f7c408a1453ff9a59e021e1fb773b3c60991d607953b9c46cad-merged.mount: Deactivated successfully. Nov 27 03:18:37 localhost podman[70252]: 2025-11-27 08:18:37.891155986 +0000 UTC m=+0.272583446 container remove 96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_taussig, ceph=True, io.openshift.tags=rhceph ceph, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, distribution-scope=public, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, vcs-type=git, maintainer=Guillaume Abrioux ) Nov 27 03:18:37 localhost systemd[1]: libpod-conmon-96c7384d804d696230020d3a9188ce31a797a589799aa79cb90d014c5190a7c5.scope: Deactivated successfully. Nov 27 03:18:38 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 03:18:38 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 03:18:38 localhost systemd[1]: man-db-cache-update.service: Consumed 1.032s CPU time. Nov 27 03:18:38 localhost systemd[1]: run-re77ea9801ba8488faadcca119cc176de.service: Deactivated successfully. Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/Package[snmpd]/ensure: created Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/File[snmpd.conf]/content: content changed '{sha256}2b743f970e80e2150759bfc66f2d8d0fbd8b31624f79e2991248d1a5ac57494e' to '{sha256}66fa6c798d40d1d9d0bd292a949bb29e7b83e5a41d66327469122442563614a6' Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/File[snmpd.sysconfig]/content: content changed '{sha256}b63afb2dee7419b6834471f88581d981c8ae5c8b27b9d329ba67a02f3ddd8221' to '{sha256}3917ee8bbc680ad50d77186ad4a1d2705c2025c32fc32f823abbda7f2328dfbd' Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/File[snmptrapd.conf]/content: content changed '{sha256}2e1ca894d609ef337b6243909bf5623c87fd5df98ecbd00c7d4c12cf12f03c4e' to '{sha256}3ecf18da1ba84ea3932607f2b903ee6a038b6f9ac4e1e371e48f3ef61c5052ea' Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/File[snmptrapd.sysconfig]/content: content changed '{sha256}86ee5797ad10cb1ea0f631e9dfa6ae278ecf4f4d16f4c80f831cdde45601b23c' to '{sha256}2244553364afcca151958f8e2003e4c182f5e2ecfbe55405cec73fd818581e97' Nov 27 03:18:38 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/Service[snmptrapd]: Triggered 'refresh' from 2 events Nov 27 03:18:41 localhost ansible-async_wrapper.py[67737]: 67738 still running (3590) Nov 27 03:18:44 localhost puppet-user[67756]: Notice: /Stage[main]/Tripleo::Profile::Base::Snmp/Snmp::Snmpv3_user[ro_snmp_user]/Exec[create-snmpv3-user-ro_snmp_user]/returns: executed successfully Nov 27 03:18:44 localhost systemd[1]: Reloading. Nov 27 03:18:44 localhost systemd-rc-local-generator[71006]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:44 localhost systemd-sysv-generator[71012]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:44 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:44 localhost systemd[1]: Starting Simple Network Management Protocol (SNMP) Daemon.... Nov 27 03:18:44 localhost snmpd[71019]: Can't find directory of RPM packages Nov 27 03:18:44 localhost snmpd[71019]: Duplicate IPv4 address detected, some interfaces may not be visible in IP-MIB Nov 27 03:18:44 localhost systemd[1]: Started Simple Network Management Protocol (SNMP) Daemon.. Nov 27 03:18:44 localhost systemd[1]: Reloading. Nov 27 03:18:44 localhost systemd-rc-local-generator[71047]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:44 localhost systemd-sysv-generator[71050]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:44 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:44 localhost systemd[1]: Reloading. Nov 27 03:18:45 localhost systemd-sysv-generator[71087]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:45 localhost systemd-rc-local-generator[71083]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:45 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:45 localhost puppet-user[67756]: Notice: /Stage[main]/Snmp/Service[snmpd]/ensure: ensure changed 'stopped' to 'running' Nov 27 03:18:45 localhost puppet-user[67756]: Notice: Applied catalog in 15.00 seconds Nov 27 03:18:45 localhost puppet-user[67756]: Application: Nov 27 03:18:45 localhost puppet-user[67756]: Initial environment: production Nov 27 03:18:45 localhost puppet-user[67756]: Converged environment: production Nov 27 03:18:45 localhost puppet-user[67756]: Run mode: user Nov 27 03:18:45 localhost puppet-user[67756]: Changes: Nov 27 03:18:45 localhost puppet-user[67756]: Total: 8 Nov 27 03:18:45 localhost puppet-user[67756]: Events: Nov 27 03:18:45 localhost puppet-user[67756]: Success: 8 Nov 27 03:18:45 localhost puppet-user[67756]: Total: 8 Nov 27 03:18:45 localhost puppet-user[67756]: Resources: Nov 27 03:18:45 localhost puppet-user[67756]: Restarted: 1 Nov 27 03:18:45 localhost puppet-user[67756]: Changed: 8 Nov 27 03:18:45 localhost puppet-user[67756]: Out of sync: 8 Nov 27 03:18:45 localhost puppet-user[67756]: Total: 19 Nov 27 03:18:45 localhost puppet-user[67756]: Time: Nov 27 03:18:45 localhost puppet-user[67756]: Filebucket: 0.00 Nov 27 03:18:45 localhost puppet-user[67756]: Schedule: 0.00 Nov 27 03:18:45 localhost puppet-user[67756]: Augeas: 0.02 Nov 27 03:18:45 localhost puppet-user[67756]: File: 0.07 Nov 27 03:18:45 localhost puppet-user[67756]: Config retrieval: 0.30 Nov 27 03:18:45 localhost puppet-user[67756]: Service: 1.22 Nov 27 03:18:45 localhost puppet-user[67756]: Transaction evaluation: 14.99 Nov 27 03:18:45 localhost puppet-user[67756]: Catalog application: 15.00 Nov 27 03:18:45 localhost puppet-user[67756]: Last run: 1764231525 Nov 27 03:18:45 localhost puppet-user[67756]: Exec: 5.05 Nov 27 03:18:45 localhost puppet-user[67756]: Package: 8.52 Nov 27 03:18:45 localhost puppet-user[67756]: Total: 15.00 Nov 27 03:18:45 localhost puppet-user[67756]: Version: Nov 27 03:18:45 localhost puppet-user[67756]: Config: 1764231509 Nov 27 03:18:45 localhost puppet-user[67756]: Puppet: 7.10.0 Nov 27 03:18:45 localhost ansible-async_wrapper.py[67738]: Module complete (67738) Nov 27 03:18:46 localhost ansible-async_wrapper.py[67737]: Done in kid B. Nov 27 03:18:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:18:47 localhost systemd[1]: tmp-crun.2PIuva.mount: Deactivated successfully. Nov 27 03:18:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:18:47 localhost podman[71109]: 2025-11-27 08:18:47.39916703 +0000 UTC m=+0.096360740 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, managed_by=tripleo_ansible, distribution-scope=public, tcib_managed=true, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, container_name=iscsid, vendor=Red Hat, Inc., config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, vcs-type=git, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:18:47 localhost podman[71109]: 2025-11-27 08:18:47.435875841 +0000 UTC m=+0.133069581 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, name=rhosp17/openstack-iscsid, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:18:47 localhost python3[71108]: ansible-ansible.legacy.async_status Invoked with jid=563485811991.67705 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:18:47 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:18:47 localhost systemd[1]: tmp-crun.L9wcXx.mount: Deactivated successfully. Nov 27 03:18:47 localhost podman[71127]: 2025-11-27 08:18:47.490540886 +0000 UTC m=+0.082158858 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, version=17.1.12, io.buildah.version=1.41.4, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, config_id=tripleo_step3, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, com.redhat.component=openstack-collectd-container, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd) Nov 27 03:18:47 localhost podman[71127]: 2025-11-27 08:18:47.529887658 +0000 UTC m=+0.121505690 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., version=17.1.12, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3) Nov 27 03:18:47 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:18:48 localhost python3[71163]: ansible-file Invoked with path=/var/lib/container-puppet/puppetlabs state=directory setype=svirt_sandbox_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:18:48 localhost python3[71179]: ansible-stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:18:49 localhost python3[71229]: ansible-ansible.legacy.stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:49 localhost python3[71247]: ansible-ansible.legacy.file Invoked with setype=svirt_sandbox_file_t selevel=s0 dest=/var/lib/container-puppet/puppetlabs/facter.conf _original_basename=tmpms4s038o recurse=False state=file path=/var/lib/container-puppet/puppetlabs/facter.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:18:49 localhost python3[71277]: ansible-file Invoked with path=/opt/puppetlabs/facter state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:50 localhost python3[71380]: ansible-ansible.posix.synchronize Invoked with src=/opt/puppetlabs/ dest=/var/lib/container-puppet/puppetlabs/ _local_rsync_path=rsync _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] ssh_connection_multiplexing=False partial=False verify_host=False mode=push dest_port=None private_key=None recursive=None links=None perms=None times=None owner=None group=None ssh_args=None link_dest=None Nov 27 03:18:51 localhost python3[71399]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:52 localhost python3[71431]: ansible-stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:18:53 localhost python3[71481]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-container-shutdown follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:53 localhost python3[71499]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-container-shutdown _original_basename=tripleo-container-shutdown recurse=False state=file path=/usr/libexec/tripleo-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:54 localhost python3[71561]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-start-podman-container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:54 localhost python3[71579]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-start-podman-container _original_basename=tripleo-start-podman-container recurse=False state=file path=/usr/libexec/tripleo-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:54 localhost python3[71641]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/tripleo-container-shutdown.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:55 localhost python3[71659]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/tripleo-container-shutdown.service _original_basename=tripleo-container-shutdown-service recurse=False state=file path=/usr/lib/systemd/system/tripleo-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:55 localhost python3[71721]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:56 localhost python3[71739]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset _original_basename=91-tripleo-container-shutdown-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:56 localhost python3[71769]: ansible-systemd Invoked with name=tripleo-container-shutdown state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:18:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:18:56 localhost systemd[1]: Reloading. Nov 27 03:18:56 localhost podman[71771]: 2025-11-27 08:18:56.739290124 +0000 UTC m=+0.081825319 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, managed_by=tripleo_ansible, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, version=17.1.12, architecture=x86_64, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, vcs-type=git) Nov 27 03:18:56 localhost systemd-rc-local-generator[71807]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:56 localhost systemd-sysv-generator[71810]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:56 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:56 localhost podman[71771]: 2025-11-27 08:18:56.920029772 +0000 UTC m=+0.262564987 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, version=17.1.12, container_name=metrics_qdr, tcib_managed=true, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:18:56 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:18:57 localhost python3[71883]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/netns-placeholder.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:57 localhost python3[71901]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/usr/lib/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:58 localhost python3[71963]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:18:58 localhost python3[71981]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:18:59 localhost python3[72011]: ansible-systemd Invoked with name=netns-placeholder state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:18:59 localhost systemd[1]: Reloading. Nov 27 03:18:59 localhost systemd-sysv-generator[72039]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:18:59 localhost systemd-rc-local-generator[72035]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:18:59 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:18:59 localhost systemd[1]: Starting Create netns directory... Nov 27 03:18:59 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 03:18:59 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 03:18:59 localhost systemd[1]: Finished Create netns directory. Nov 27 03:19:00 localhost python3[72068]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 03:19:01 localhost python3[72126]: ansible-tripleo_container_manage Invoked with config_id=tripleo_step4 config_dir=/var/lib/tripleo-config/container-startup-config/step_4 config_patterns=*.json config_overrides={} concurrency=5 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.187474538 +0000 UTC m=+0.086755232 container create 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, architecture=x86_64, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, container_name=configure_cms_options, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, batch=17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, release=1761123044) Nov 27 03:19:02 localhost systemd[1]: Started libpod-conmon-8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c.scope. Nov 27 03:19:02 localhost podman[72292]: 2025-11-27 08:19:02.222235255 +0000 UTC m=+0.093435522 container create e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, distribution-scope=public, tcib_managed=true, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4, container_name=logrotate_crond, release=1761123044, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, url=https://www.redhat.com, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z) Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.138002662 +0000 UTC m=+0.037283416 image pull registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:19:02 localhost systemd[1]: Started libcrun container. Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.246344586 +0000 UTC m=+0.113151204 container create 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., container_name=nova_libvirt_init_secret, version=17.1.12, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-type=git, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z) Nov 27 03:19:02 localhost systemd[1]: Started libpod-conmon-e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.scope. Nov 27 03:19:02 localhost podman[72319]: 2025-11-27 08:19:02.259348147 +0000 UTC m=+0.094414959 container create ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, config_id=tripleo_step4, io.openshift.expose-services=, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.262455891 +0000 UTC m=+0.161736585 container init 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, release=1761123044, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, url=https://www.redhat.com, batch=17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, config_id=tripleo_step4, container_name=configure_cms_options, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ovn-controller-container, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:19:02 localhost podman[72292]: 2025-11-27 08:19:02.163090889 +0000 UTC m=+0.034291166 image pull registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.270708814 +0000 UTC m=+0.169989508 container start 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, managed_by=tripleo_ansible, io.buildah.version=1.41.4, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_id=tripleo_step4, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, architecture=x86_64, vendor=Red Hat, Inc., tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=configure_cms_options, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.271278569 +0000 UTC m=+0.170559263 container attach 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, container_name=configure_cms_options, distribution-scope=public, url=https://www.redhat.com, version=17.1.12, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-type=git) Nov 27 03:19:02 localhost systemd[1]: Started libcrun container. Nov 27 03:19:02 localhost systemd[1]: Started libpod-conmon-7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8.scope. Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.180643274 +0000 UTC m=+0.047449882 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6fdd4104eccc8c43c3ebaef353a49eae132b42626593a557e6485db08bad4f8e/merged/var/log/containers supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost systemd[1]: Started libcrun container. Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/085ba0108fbd311421129957b0cab393dd105989700ee149ddb4c3f62f0d2357/merged/etc/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/085ba0108fbd311421129957b0cab393dd105989700ee149ddb4c3f62f0d2357/merged/etc/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/085ba0108fbd311421129957b0cab393dd105989700ee149ddb4c3f62f0d2357/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.297234419 +0000 UTC m=+0.164041037 container init 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-libvirt, architecture=x86_64, build-date=2025-11-19T00:35:22Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, com.redhat.component=openstack-nova-libvirt-container, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, container_name=nova_libvirt_init_secret) Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.30280762 +0000 UTC m=+0.169614218 container start 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_libvirt_init_secret, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, build-date=2025-11-19T00:35:22Z, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, tcib_managed=true, batch=17.1_20251118.1, distribution-scope=public, release=1761123044, name=rhosp17/openstack-nova-libvirt, vendor=Red Hat, Inc., io.openshift.expose-services=, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.303221031 +0000 UTC m=+0.170027669 container attach 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-libvirt-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, tcib_managed=true, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, release=1761123044, name=rhosp17/openstack-nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., container_name=nova_libvirt_init_secret, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, build-date=2025-11-19T00:35:22Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, config_id=tripleo_step4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4) Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:19:02 localhost podman[72292]: 2025-11-27 08:19:02.306910321 +0000 UTC m=+0.178110578 container init e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, build-date=2025-11-18T22:49:32Z, vcs-type=git, config_id=tripleo_step4, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.expose-services=, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, tcib_managed=true) Nov 27 03:19:02 localhost podman[72319]: 2025-11-27 08:19:02.214194838 +0000 UTC m=+0.049261650 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:19:02 localhost podman[72292]: 2025-11-27 08:19:02.333600921 +0000 UTC m=+0.204801168 container start e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, managed_by=tripleo_ansible, vcs-type=git, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, build-date=2025-11-18T22:49:32Z) Nov 27 03:19:02 localhost systemd[1]: Started libpod-conmon-ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.scope. Nov 27 03:19:02 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name logrotate_crond --conmon-pidfile /run/logrotate_crond.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=53ed83bb0cae779ff95edb2002262c6f --healthcheck-command /usr/share/openstack-tripleo-common/healthcheck/cron --label config_id=tripleo_step4 --label container_name=logrotate_crond --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/logrotate_crond.log --network none --pid host --privileged=True --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro --volume /var/log/containers:/var/log/containers:z registry.redhat.io/rhosp-rhel9/openstack-cron:17.1 Nov 27 03:19:02 localhost systemd[1]: Started libcrun container. Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/e7f27f4aecc7722c6029129a1910e39be572d823c20ebf6e60c902dacc886f74/merged/var/log/ceilometer supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost ovs-vsctl[72409]: ovs|00001|vsctl|INFO|Called as ovs-vsctl remove open . external_ids ovn-cms-options Nov 27 03:19:02 localhost podman[72263]: 2025-11-27 08:19:02.400007494 +0000 UTC m=+0.299288208 container died 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=configure_cms_options, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, tcib_managed=true, config_id=tripleo_step4, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, version=17.1.12, vendor=Red Hat, Inc., managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-type=git, batch=17.1_20251118.1) Nov 27 03:19:02 localhost systemd[1]: libpod-8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c.scope: Deactivated successfully. Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:19:02 localhost podman[72319]: 2025-11-27 08:19:02.409863669 +0000 UTC m=+0.244930491 container init ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, config_id=tripleo_step4, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vendor=Red Hat, Inc., batch=17.1_20251118.1, release=1761123044) Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:19:02 localhost podman[72319]: 2025-11-27 08:19:02.427724861 +0000 UTC m=+0.262791683 container start ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, url=https://www.redhat.com, tcib_managed=true, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-type=git, batch=17.1_20251118.1, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:19:02 localhost podman[72350]: 2025-11-27 08:19:02.328042411 +0000 UTC m=+0.105982161 image pull registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 Nov 27 03:19:02 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name ceilometer_agent_ipmi --conmon-pidfile /run/ceilometer_agent_ipmi.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=d7f975e76c3815f85351b6258d097495 --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step4 --label container_name=ceilometer_agent_ipmi --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/ceilometer_agent_ipmi.log --network host --privileged=True --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro --volume /var/log/containers/ceilometer:/var/log/ceilometer:z registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1 Nov 27 03:19:02 localhost systemd[1]: libpod-7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8.scope: Deactivated successfully. Nov 27 03:19:02 localhost podman[72432]: 2025-11-27 08:19:02.488050558 +0000 UTC m=+0.057845561 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=starting, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, distribution-scope=public, tcib_managed=true, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, batch=17.1_20251118.1, version=17.1.12, architecture=x86_64, release=1761123044, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi) Nov 27 03:19:02 localhost podman[72350]: 2025-11-27 08:19:02.525490749 +0000 UTC m=+0.303430489 container create f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., io.openshift.expose-services=, tcib_managed=true, vcs-type=git, build-date=2025-11-19T00:11:48Z, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, url=https://www.redhat.com, config_id=tripleo_step4, version=17.1.12, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:19:02 localhost systemd[1]: Started libpod-conmon-f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.scope. Nov 27 03:19:02 localhost podman[72432]: 2025-11-27 08:19:02.572462986 +0000 UTC m=+0.142257989 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, version=17.1.12, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:19:02 localhost systemd[1]: Started libcrun container. Nov 27 03:19:02 localhost podman[72432]: unhealthy Nov 27 03:19:02 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:19:02 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 03:19:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3ed13769fb4ebcfcdb2db8fb9f0b5f0e5ef510466865cb9e3869aeb9bf07dd72/merged/var/log/ceilometer supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:02 localhost podman[72413]: 2025-11-27 08:19:02.627586744 +0000 UTC m=+0.221996921 container cleanup 8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=configure_cms_options, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=configure_cms_options, version=17.1.12, batch=17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., build-date=2025-11-18T23:34:05Z, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:19:02 localhost systemd[1]: libpod-conmon-8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c.scope: Deactivated successfully. Nov 27 03:19:02 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name configure_cms_options --conmon-pidfile /run/configure_cms_options.pid --detach=False --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --label config_id=tripleo_step4 --label container_name=configure_cms_options --label managed_by=tripleo_ansible --label config_data={'command': ['/bin/bash', '-c', 'CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/configure_cms_options.log --network host --privileged=True --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /lib/modules:/lib/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 /bin/bash -c CMS_OPTS=$(hiera ovn::controller::ovn_cms_options -c /etc/puppet/hiera.yaml); if [ X"$CMS_OPTS" != X ]; then ovs-vsctl set open . external_ids:ovn-cms-options=$CMS_OPTS;else ovs-vsctl remove open . external_ids ovn-cms-options; fi Nov 27 03:19:02 localhost podman[72390]: 2025-11-27 08:19:02.441658027 +0000 UTC m=+0.099991760 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=starting, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-type=git, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, release=1761123044, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:19:02 localhost podman[72350]: 2025-11-27 08:19:02.667079959 +0000 UTC m=+0.445019719 container init f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, container_name=ceilometer_agent_compute, config_id=tripleo_step4, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, release=1761123044, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:19:02 localhost podman[72390]: 2025-11-27 08:19:02.671428157 +0000 UTC m=+0.329761900 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true, architecture=x86_64, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public) Nov 27 03:19:02 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:19:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:19:02 localhost podman[72350]: 2025-11-27 08:19:02.703751759 +0000 UTC m=+0.481691519 container start f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, container_name=ceilometer_agent_compute, managed_by=tripleo_ansible, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, distribution-scope=public, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, release=1761123044, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4) Nov 27 03:19:02 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name ceilometer_agent_compute --conmon-pidfile /run/ceilometer_agent_compute.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=d7f975e76c3815f85351b6258d097495 --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step4 --label container_name=ceilometer_agent_compute --label managed_by=tripleo_ansible --label config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/ceilometer_agent_compute.log --network host --privileged=False --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/log/containers/ceilometer:/var/log/ceilometer:z registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1 Nov 27 03:19:02 localhost podman[72289]: 2025-11-27 08:19:02.725819994 +0000 UTC m=+0.592626622 container died 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, build-date=2025-11-19T00:35:22Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-libvirt-container, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vendor=Red Hat, Inc., container_name=nova_libvirt_init_secret, vcs-type=git, name=rhosp17/openstack-nova-libvirt, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:19:02 localhost podman[72467]: 2025-11-27 08:19:02.759160734 +0000 UTC m=+0.280454789 container cleanup 7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_libvirt_init_secret, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, name=rhosp17/openstack-nova-libvirt, distribution-scope=public, url=https://www.redhat.com, version=17.1.12, tcib_managed=true, vcs-type=git, config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']}, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-libvirt-container, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_libvirt_init_secret, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, release=1761123044, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, build-date=2025-11-19T00:35:22Z) Nov 27 03:19:02 localhost systemd[1]: libpod-conmon-7432c8eb24a4aa1d7699ff5055161db13204107a643b80a184afee1fb60742e8.scope: Deactivated successfully. Nov 27 03:19:02 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_libvirt_init_secret --cgroupns=host --conmon-pidfile /run/nova_libvirt_init_secret.pid --detach=False --env LIBVIRT_DEFAULT_URI=qemu:///system --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step4 --label container_name=nova_libvirt_init_secret --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'command': '/nova_libvirt_init_secret.sh ceph:openstack', 'detach': False, 'environment': {'LIBVIRT_DEFAULT_URI': 'qemu:///system', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'privileged': False, 'security_opt': ['label=disable'], 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova', '/etc/libvirt:/etc/libvirt', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro', '/var/lib/tripleo-config/ceph:/etc/ceph:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_libvirt_init_secret.log --network host --privileged=False --security-opt label=disable --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/config-data/puppet-generated/nova_libvirt/etc/nova:/etc/nova --volume /etc/libvirt:/etc/libvirt --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /var/lib/container-config-scripts/nova_libvirt_init_secret.sh:/nova_libvirt_init_secret.sh:ro --volume /var/lib/tripleo-config/ceph:/etc/ceph:ro registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1 /nova_libvirt_init_secret.sh ceph:openstack Nov 27 03:19:02 localhost podman[72543]: 2025-11-27 08:19:02.815678189 +0000 UTC m=+0.111974123 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=starting, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, io.openshift.expose-services=, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:19:02 localhost podman[72543]: 2025-11-27 08:19:02.907167398 +0000 UTC m=+0.203463322 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, url=https://www.redhat.com, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, build-date=2025-11-19T00:11:48Z, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, version=17.1.12, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:19:02 localhost podman[72543]: unhealthy Nov 27 03:19:02 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:19:02 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed with result 'exit-code'. Nov 27 03:19:03 localhost podman[72677]: 2025-11-27 08:19:03.082849589 +0000 UTC m=+0.100973365 container create f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, vendor=Red Hat, Inc., architecture=x86_64, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, container_name=nova_migration_target, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:19:03 localhost podman[72677]: 2025-11-27 08:19:03.032066968 +0000 UTC m=+0.050190814 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:19:03 localhost podman[72698]: 2025-11-27 08:19:03.132834397 +0000 UTC m=+0.093715539 container create e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, container_name=setup_ovs_manager, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, distribution-scope=public, architecture=x86_64, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, batch=17.1_20251118.1, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, vcs-type=git, version=17.1.12, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:19:03 localhost systemd[1]: Started libpod-conmon-e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27.scope. Nov 27 03:19:03 localhost systemd[1]: Started libcrun container. Nov 27 03:19:03 localhost podman[72698]: 2025-11-27 08:19:03.091226055 +0000 UTC m=+0.052107247 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 03:19:03 localhost systemd[1]: var-lib-containers-storage-overlay-a149a1b5b441534a5a876cad2d9a16ef9b7b119133a8f5528aa76afae576e593-merged.mount: Deactivated successfully. Nov 27 03:19:03 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-8df3bcf0bb3c1b2a450eb8c4f03f09ff463737b942246fa93ef31cb56b450c9c-userdata-shm.mount: Deactivated successfully. Nov 27 03:19:03 localhost podman[72698]: 2025-11-27 08:19:03.225642112 +0000 UTC m=+0.186523254 container init e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, version=17.1.12, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=setup_ovs_manager, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, io.openshift.expose-services=, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com) Nov 27 03:19:03 localhost systemd[1]: Started libpod-conmon-f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.scope. Nov 27 03:19:03 localhost podman[72698]: 2025-11-27 08:19:03.235977951 +0000 UTC m=+0.196859063 container start e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, release=1761123044, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=setup_ovs_manager, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, batch=17.1_20251118.1) Nov 27 03:19:03 localhost podman[72698]: 2025-11-27 08:19:03.236202717 +0000 UTC m=+0.197083899 container attach e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, distribution-scope=public, batch=17.1_20251118.1, vcs-type=git, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, architecture=x86_64, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, container_name=setup_ovs_manager, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:19:03 localhost systemd[1]: Started libcrun container. Nov 27 03:19:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f46c77aa3acfba4abd8d96e809917692cad69270a1f593d1d2c9063645aac7e1/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:19:03 localhost podman[72677]: 2025-11-27 08:19:03.279168177 +0000 UTC m=+0.297291963 container init f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, vendor=Red Hat, Inc., architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, vcs-type=git, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.openshift.expose-services=) Nov 27 03:19:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:19:03 localhost podman[72677]: 2025-11-27 08:19:03.319839724 +0000 UTC m=+0.337963520 container start f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64) Nov 27 03:19:03 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_migration_target --conmon-pidfile /run/nova_migration_target.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step4 --label container_name=nova_migration_target --label managed_by=tripleo_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_migration_target.log --network host --privileged=True --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /etc/ssh:/host-ssh:ro --volume /run/libvirt:/run/libvirt:shared,z --volume /var/lib/nova:/var/lib/nova:shared registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:19:03 localhost podman[72732]: 2025-11-27 08:19:03.410515631 +0000 UTC m=+0.083138595 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=starting, container_name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, batch=17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., tcib_managed=true, config_id=tripleo_step4, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:19:03 localhost podman[72732]: 2025-11-27 08:19:03.753884726 +0000 UTC m=+0.426507690 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, container_name=nova_migration_target, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, distribution-scope=public) Nov 27 03:19:03 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:19:03 localhost kernel: capability: warning: `privsep-helper' uses deprecated v2 capabilities in a way that may be insecure Nov 27 03:19:05 localhost ovs-vsctl[72906]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --timeout=5 --id=@manager -- create Manager "target=\"ptcp:6640:127.0.0.1\"" -- add Open_vSwitch . manager_options @manager Nov 27 03:19:06 localhost systemd[1]: libpod-e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27.scope: Deactivated successfully. Nov 27 03:19:06 localhost systemd[1]: libpod-e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27.scope: Consumed 2.802s CPU time. Nov 27 03:19:06 localhost podman[72907]: 2025-11-27 08:19:06.117308791 +0000 UTC m=+0.045473518 container died e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, container_name=setup_ovs_manager, vendor=Red Hat, Inc., io.buildah.version=1.41.4, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, config_id=tripleo_step4, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:19:06 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27-userdata-shm.mount: Deactivated successfully. Nov 27 03:19:06 localhost systemd[1]: var-lib-containers-storage-overlay-4355f530465b936c6835ef8329885a7e2b3c0f1075ad1c8dfe83795d5120a50b-merged.mount: Deactivated successfully. Nov 27 03:19:06 localhost podman[72907]: 2025-11-27 08:19:06.159321405 +0000 UTC m=+0.087486142 container cleanup e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=setup_ovs_manager, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=setup_ovs_manager, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, batch=17.1_20251118.1, config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, release=1761123044, config_id=tripleo_step4, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:19:06 localhost systemd[1]: libpod-conmon-e74e65862730b1fda2a7ca7cdb4e9a896dd5a9f49f66079ebbf9f3a7815caf27.scope: Deactivated successfully. Nov 27 03:19:06 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name setup_ovs_manager --conmon-pidfile /run/setup_ovs_manager.pid --detach=False --env TRIPLEO_DEPLOY_IDENTIFIER=1764229719 --label config_id=tripleo_step4 --label container_name=setup_ovs_manager --label managed_by=tripleo_ansible --label config_data={'command': ['/container_puppet_apply.sh', '4', 'exec', 'include tripleo::profile::base::neutron::ovn_metadata'], 'detach': False, 'environment': {'TRIPLEO_DEPLOY_IDENTIFIER': '1764229719'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'privileged': True, 'start_order': 0, 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro', '/etc/puppet:/tmp/puppet-etc:ro', '/usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/setup_ovs_manager.log --network host --privileged=True --user root --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /var/lib/container-config-scripts/container_puppet_apply.sh:/container_puppet_apply.sh:ro --volume /etc/puppet:/tmp/puppet-etc:ro --volume /usr/share/openstack-puppet/modules:/usr/share/openstack-puppet/modules:ro --volume /lib/modules:/lib/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 /container_puppet_apply.sh 4 exec include tripleo::profile::base::neutron::ovn_metadata Nov 27 03:19:06 localhost podman[73010]: 2025-11-27 08:19:06.632592186 +0000 UTC m=+0.082889089 container create 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, container_name=ovn_metadata_agent, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, managed_by=tripleo_ansible, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:19:06 localhost systemd[1]: Started libpod-conmon-2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.scope. Nov 27 03:19:06 localhost podman[73010]: 2025-11-27 08:19:06.587055706 +0000 UTC m=+0.037352649 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 03:19:06 localhost podman[73031]: 2025-11-27 08:19:06.698491873 +0000 UTC m=+0.096768061 container create aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, version=17.1.12, managed_by=tripleo_ansible, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z) Nov 27 03:19:06 localhost systemd[1]: Started libcrun container. Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f96fef7b313009df8cfca80855d5a7b323e2e769d6b440ee9ef0691972acd2bc/merged/etc/neutron/kill_scripts supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f96fef7b313009df8cfca80855d5a7b323e2e769d6b440ee9ef0691972acd2bc/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f96fef7b313009df8cfca80855d5a7b323e2e769d6b440ee9ef0691972acd2bc/merged/var/log/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost systemd[1]: Started libpod-conmon-aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.scope. Nov 27 03:19:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:19:06 localhost podman[73010]: 2025-11-27 08:19:06.747005803 +0000 UTC m=+0.197302706 container init 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, vendor=Red Hat, Inc., version=17.1.12, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, tcib_managed=true, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 03:19:06 localhost systemd[1]: Started libcrun container. Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6cf7f72aaf45e45d932f6befc623c64b0e80d1af7cd41be6a403137845b9774f/merged/run/ovn supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6cf7f72aaf45e45d932f6befc623c64b0e80d1af7cd41be6a403137845b9774f/merged/var/log/ovn supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6cf7f72aaf45e45d932f6befc623c64b0e80d1af7cd41be6a403137845b9774f/merged/var/log/openvswitch supports timestamps until 2038 (0x7fffffff) Nov 27 03:19:06 localhost podman[73031]: 2025-11-27 08:19:06.655943875 +0000 UTC m=+0.054219963 image pull registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:19:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:19:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:19:06 localhost podman[73031]: 2025-11-27 08:19:06.782936262 +0000 UTC m=+0.181212350 container init aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, release=1761123044, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, name=rhosp17/openstack-ovn-controller, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, container_name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4) Nov 27 03:19:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:19:06 localhost podman[73031]: 2025-11-27 08:19:06.811718709 +0000 UTC m=+0.209994767 container start aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, build-date=2025-11-18T23:34:05Z, io.buildah.version=1.41.4, vcs-type=git, release=1761123044, architecture=x86_64, managed_by=tripleo_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, version=17.1.12, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:19:06 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:19:06 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name ovn_controller --conmon-pidfile /run/ovn_controller.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --healthcheck-command /openstack/healthcheck 6642 --label config_id=tripleo_step4 --label container_name=ovn_controller --label managed_by=tripleo_ansible --label config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/ovn_controller.log --network host --privileged=True --user root --volume /var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro --volume /lib/modules:/lib/modules:ro --volume /run:/run --volume /var/lib/openvswitch/ovn:/run/ovn:shared,z --volume /var/log/containers/openvswitch:/var/log/openvswitch:z --volume /var/log/containers/openvswitch:/var/log/ovn:z registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1 Nov 27 03:19:06 localhost podman[73010]: 2025-11-27 08:19:06.815622684 +0000 UTC m=+0.265919567 container start 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, io.buildah.version=1.41.4, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, version=17.1.12, vcs-type=git, release=1761123044, tcib_managed=true, container_name=ovn_metadata_agent, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=) Nov 27 03:19:06 localhost systemd[1]: Created slice User Slice of UID 0. Nov 27 03:19:06 localhost systemd[1]: Starting User Runtime Directory /run/user/0... Nov 27 03:19:06 localhost python3[72126]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name ovn_metadata_agent --cgroupns=host --conmon-pidfile /run/ovn_metadata_agent.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env TRIPLEO_CONFIG_HASH=8cac2334de855b5d814608ea6483099b --healthcheck-command /openstack/healthcheck --label config_id=tripleo_step4 --label container_name=ovn_metadata_agent --label managed_by=tripleo_ansible --label config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/ovn_metadata_agent.log --network host --pid host --privileged=True --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/neutron:/var/log/neutron:z --volume /var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro --volume /lib/modules:/lib/modules:ro --volume /run/openvswitch:/run/openvswitch:shared,z --volume /var/lib/neutron:/var/lib/neutron:shared,z --volume /run/netns:/run/netns:shared --volume /var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z --volume /var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 03:19:06 localhost podman[73058]: 2025-11-27 08:19:06.855354126 +0000 UTC m=+0.081101720 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=starting, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, tcib_managed=true, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, architecture=x86_64, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, version=17.1.12, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn) Nov 27 03:19:06 localhost systemd[1]: Finished User Runtime Directory /run/user/0. Nov 27 03:19:06 localhost systemd[1]: Starting User Manager for UID 0... Nov 27 03:19:06 localhost podman[73058]: 2025-11-27 08:19:06.898043498 +0000 UTC m=+0.123791092 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, config_id=tripleo_step4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, batch=17.1_20251118.1, release=1761123044, architecture=x86_64, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:19:06 localhost podman[73058]: unhealthy Nov 27 03:19:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:19:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:19:06 localhost podman[73074]: 2025-11-27 08:19:06.922650502 +0000 UTC m=+0.103768461 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=starting, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, url=https://www.redhat.com, container_name=ovn_controller, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4) Nov 27 03:19:07 localhost podman[73074]: 2025-11-27 08:19:07.006868674 +0000 UTC m=+0.187986683 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, container_name=ovn_controller, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, config_id=tripleo_step4, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, batch=17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:19:07 localhost podman[73074]: unhealthy Nov 27 03:19:07 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:19:07 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:19:07 localhost systemd[73104]: Queued start job for default target Main User Target. Nov 27 03:19:07 localhost systemd[73104]: Created slice User Application Slice. Nov 27 03:19:07 localhost systemd[73104]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). Nov 27 03:19:07 localhost systemd[73104]: Started Daily Cleanup of User's Temporary Directories. Nov 27 03:19:07 localhost systemd[73104]: Reached target Paths. Nov 27 03:19:07 localhost systemd[73104]: Reached target Timers. Nov 27 03:19:07 localhost systemd[73104]: Starting D-Bus User Message Bus Socket... Nov 27 03:19:07 localhost systemd[73104]: Starting Create User's Volatile Files and Directories... Nov 27 03:19:07 localhost systemd[73104]: Listening on D-Bus User Message Bus Socket. Nov 27 03:19:07 localhost systemd[73104]: Reached target Sockets. Nov 27 03:19:07 localhost systemd[73104]: Finished Create User's Volatile Files and Directories. Nov 27 03:19:07 localhost systemd[73104]: Reached target Basic System. Nov 27 03:19:07 localhost systemd[73104]: Reached target Main User Target. Nov 27 03:19:07 localhost systemd[73104]: Startup finished in 167ms. Nov 27 03:19:07 localhost systemd[1]: Started User Manager for UID 0. Nov 27 03:19:07 localhost systemd[1]: Started Session c9 of User root. Nov 27 03:19:07 localhost systemd[1]: session-c9.scope: Deactivated successfully. Nov 27 03:19:07 localhost kernel: device br-int entered promiscuous mode Nov 27 03:19:07 localhost NetworkManager[5971]: [1764231547.1609] manager: (br-int): new Generic device (/org/freedesktop/NetworkManager/Devices/11) Nov 27 03:19:07 localhost systemd-udevd[73175]: Network interface NamePolicy= disabled on kernel command line. Nov 27 03:19:07 localhost python3[73195]: ansible-file Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_compute.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:07 localhost python3[73211]: ansible-file Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_ipmi.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:08 localhost python3[73227]: ansible-file Invoked with path=/etc/systemd/system/tripleo_logrotate_crond.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:08 localhost python3[73243]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_migration_target.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:08 localhost kernel: device genev_sys_6081 entered promiscuous mode Nov 27 03:19:08 localhost NetworkManager[5971]: [1764231548.2482] device (genev_sys_6081): carrier: link connected Nov 27 03:19:08 localhost NetworkManager[5971]: [1764231548.2484] manager: (genev_sys_6081): new Generic device (/org/freedesktop/NetworkManager/Devices/12) Nov 27 03:19:08 localhost python3[73265]: ansible-file Invoked with path=/etc/systemd/system/tripleo_ovn_controller.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:08 localhost python3[73284]: ansible-file Invoked with path=/etc/systemd/system/tripleo_ovn_metadata_agent.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:08 localhost python3[73301]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_compute_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:09 localhost python3[73318]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_ipmi_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:09 localhost python3[73335]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_logrotate_crond_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:09 localhost python3[73353]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_migration_target_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:09 localhost python3[73369]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_ovn_controller_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:10 localhost python3[73385]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_ovn_metadata_agent_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:19:10 localhost python3[73446]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_ceilometer_agent_compute.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:11 localhost python3[73475]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_ceilometer_agent_ipmi.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:11 localhost python3[73504]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_logrotate_crond.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:12 localhost python3[73533]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_nova_migration_target.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:13 localhost python3[73562]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_ovn_controller.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:13 localhost python3[73591]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231550.2955878-109742-114842100799953/source dest=/etc/systemd/system/tripleo_ovn_metadata_agent.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:13 localhost python3[73607]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 03:19:13 localhost systemd[1]: Reloading. Nov 27 03:19:14 localhost systemd-sysv-generator[73633]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:14 localhost systemd-rc-local-generator[73628]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:14 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:14 localhost python3[73659]: ansible-systemd Invoked with state=restarted name=tripleo_ceilometer_agent_compute.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:14 localhost systemd[1]: Reloading. Nov 27 03:19:15 localhost systemd-rc-local-generator[73686]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:15 localhost systemd-sysv-generator[73692]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:15 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:15 localhost systemd[1]: Starting ceilometer_agent_compute container... Nov 27 03:19:15 localhost tripleo-start-podman-container[73699]: Creating additional drop-in dependency for "ceilometer_agent_compute" (f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a) Nov 27 03:19:15 localhost systemd[1]: Reloading. Nov 27 03:19:15 localhost systemd-rc-local-generator[73757]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:15 localhost systemd-sysv-generator[73762]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:15 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:15 localhost systemd[1]: Started ceilometer_agent_compute container. Nov 27 03:19:16 localhost python3[73785]: ansible-systemd Invoked with state=restarted name=tripleo_ceilometer_agent_ipmi.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:16 localhost systemd[1]: Reloading. Nov 27 03:19:16 localhost systemd-rc-local-generator[73810]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:16 localhost systemd-sysv-generator[73814]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:16 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:17 localhost systemd[1]: Starting ceilometer_agent_ipmi container... Nov 27 03:19:17 localhost systemd[1]: Started ceilometer_agent_ipmi container. Nov 27 03:19:17 localhost systemd[1]: Stopping User Manager for UID 0... Nov 27 03:19:17 localhost systemd[73104]: Activating special unit Exit the Session... Nov 27 03:19:17 localhost systemd[73104]: Stopped target Main User Target. Nov 27 03:19:17 localhost systemd[73104]: Stopped target Basic System. Nov 27 03:19:17 localhost systemd[73104]: Stopped target Paths. Nov 27 03:19:17 localhost systemd[73104]: Stopped target Sockets. Nov 27 03:19:17 localhost systemd[73104]: Stopped target Timers. Nov 27 03:19:17 localhost systemd[73104]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 03:19:17 localhost systemd[73104]: Closed D-Bus User Message Bus Socket. Nov 27 03:19:17 localhost systemd[73104]: Stopped Create User's Volatile Files and Directories. Nov 27 03:19:17 localhost systemd[73104]: Removed slice User Application Slice. Nov 27 03:19:17 localhost systemd[73104]: Reached target Shutdown. Nov 27 03:19:17 localhost systemd[73104]: Finished Exit the Session. Nov 27 03:19:17 localhost systemd[73104]: Reached target Exit the Session. Nov 27 03:19:17 localhost systemd[1]: user@0.service: Deactivated successfully. Nov 27 03:19:17 localhost systemd[1]: Stopped User Manager for UID 0. Nov 27 03:19:17 localhost systemd[1]: Stopping User Runtime Directory /run/user/0... Nov 27 03:19:17 localhost systemd[1]: run-user-0.mount: Deactivated successfully. Nov 27 03:19:17 localhost systemd[1]: user-runtime-dir@0.service: Deactivated successfully. Nov 27 03:19:17 localhost systemd[1]: Stopped User Runtime Directory /run/user/0. Nov 27 03:19:17 localhost systemd[1]: Removed slice User Slice of UID 0. Nov 27 03:19:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:19:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:19:17 localhost systemd[1]: tmp-crun.OlCKnP.mount: Deactivated successfully. Nov 27 03:19:17 localhost podman[73854]: 2025-11-27 08:19:17.643398751 +0000 UTC m=+0.091567751 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, tcib_managed=true, distribution-scope=public, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=iscsid, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com) Nov 27 03:19:17 localhost podman[73854]: 2025-11-27 08:19:17.685870017 +0000 UTC m=+0.134039007 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, tcib_managed=true, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, managed_by=tripleo_ansible, release=1761123044) Nov 27 03:19:17 localhost podman[73871]: 2025-11-27 08:19:17.72339421 +0000 UTC m=+0.076106694 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., batch=17.1_20251118.1, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, tcib_managed=true, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, vcs-type=git, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12) Nov 27 03:19:17 localhost podman[73871]: 2025-11-27 08:19:17.735886727 +0000 UTC m=+0.088599201 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-collectd, config_id=tripleo_step3, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=) Nov 27 03:19:17 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:19:17 localhost python3[73853]: ansible-systemd Invoked with state=restarted name=tripleo_logrotate_crond.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:17 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:19:17 localhost systemd[1]: Reloading. Nov 27 03:19:17 localhost systemd-rc-local-generator[73917]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:17 localhost systemd-sysv-generator[73921]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:18 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:18 localhost systemd[1]: Starting logrotate_crond container... Nov 27 03:19:18 localhost systemd[1]: Started logrotate_crond container. Nov 27 03:19:18 localhost python3[73960]: ansible-systemd Invoked with state=restarted name=tripleo_nova_migration_target.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:20 localhost systemd[1]: Reloading. Nov 27 03:19:20 localhost systemd-rc-local-generator[73985]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:20 localhost systemd-sysv-generator[73989]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:20 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:20 localhost systemd[1]: Starting nova_migration_target container... Nov 27 03:19:20 localhost systemd[1]: Started nova_migration_target container. Nov 27 03:19:21 localhost python3[74026]: ansible-systemd Invoked with state=restarted name=tripleo_ovn_controller.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:21 localhost systemd[1]: Reloading. Nov 27 03:19:21 localhost systemd-rc-local-generator[74053]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:21 localhost systemd-sysv-generator[74057]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:21 localhost systemd[1]: Starting ovn_controller container... Nov 27 03:19:21 localhost tripleo-start-podman-container[74065]: Creating additional drop-in dependency for "ovn_controller" (aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2) Nov 27 03:19:22 localhost systemd[1]: Reloading. Nov 27 03:19:22 localhost systemd-rc-local-generator[74122]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:22 localhost systemd-sysv-generator[74127]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:22 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:22 localhost systemd[1]: Started ovn_controller container. Nov 27 03:19:22 localhost python3[74149]: ansible-systemd Invoked with state=restarted name=tripleo_ovn_metadata_agent.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:19:24 localhost systemd[1]: Reloading. Nov 27 03:19:24 localhost systemd-rc-local-generator[74176]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:19:24 localhost systemd-sysv-generator[74180]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:19:24 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:19:24 localhost systemd[1]: Starting ovn_metadata_agent container... Nov 27 03:19:24 localhost systemd[1]: Started ovn_metadata_agent container. Nov 27 03:19:24 localhost python3[74230]: ansible-file Invoked with path=/var/lib/container-puppet/container-puppet-tasks4.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:26 localhost python3[74352]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=True puppet_config=/var/lib/container-puppet/container-puppet-tasks4.json short_hostname=np0005537446 step=4 update_config_hash_only=False Nov 27 03:19:26 localhost python3[74368]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:19:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:19:27 localhost systemd[1]: tmp-crun.gD5FAA.mount: Deactivated successfully. Nov 27 03:19:27 localhost podman[74385]: 2025-11-27 08:19:27.31280654 +0000 UTC m=+0.098527829 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, tcib_managed=true, container_name=metrics_qdr, vendor=Red Hat, Inc., architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, release=1761123044, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:19:27 localhost python3[74384]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_4 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:19:27 localhost podman[74385]: 2025-11-27 08:19:27.51661633 +0000 UTC m=+0.302337559 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:19:27 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:19:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:19:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:19:32 localhost podman[74417]: 2025-11-27 08:19:32.980581709 +0000 UTC m=+0.074005827 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, name=rhosp17/openstack-cron, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, tcib_managed=true, io.openshift.expose-services=, managed_by=tripleo_ansible, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044) Nov 27 03:19:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:19:33 localhost podman[74417]: 2025-11-27 08:19:33.02507515 +0000 UTC m=+0.118499268 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, config_id=tripleo_step4, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, container_name=logrotate_crond, distribution-scope=public, managed_by=tripleo_ansible, tcib_managed=true, name=rhosp17/openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc.) Nov 27 03:19:33 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:19:33 localhost podman[74416]: 2025-11-27 08:19:33.035881272 +0000 UTC m=+0.131998373 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=starting, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, version=17.1.12, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:19:33 localhost podman[74447]: 2025-11-27 08:19:33.134921704 +0000 UTC m=+0.128548290 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=starting, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, distribution-scope=public) Nov 27 03:19:33 localhost podman[74416]: 2025-11-27 08:19:33.153434294 +0000 UTC m=+0.249551355 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, version=17.1.12, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, url=https://www.redhat.com, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:19:33 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:19:33 localhost podman[74447]: 2025-11-27 08:19:33.191930023 +0000 UTC m=+0.185556609 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, managed_by=tripleo_ansible, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, release=1761123044, distribution-scope=public, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, url=https://www.redhat.com, io.openshift.expose-services=, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, version=17.1.12, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:19:33 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:19:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:19:33 localhost systemd[1]: tmp-crun.EOxXTM.mount: Deactivated successfully. Nov 27 03:19:33 localhost podman[74489]: 2025-11-27 08:19:33.989211346 +0000 UTC m=+0.088866349 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, architecture=x86_64, container_name=nova_migration_target, managed_by=tripleo_ansible, vcs-type=git, url=https://www.redhat.com, io.openshift.expose-services=, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team) Nov 27 03:19:34 localhost podman[74489]: 2025-11-27 08:19:34.330981508 +0000 UTC m=+0.430636491 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., io.buildah.version=1.41.4, container_name=nova_migration_target, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, vcs-type=git, url=https://www.redhat.com) Nov 27 03:19:34 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:19:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:19:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:19:37 localhost systemd[1]: tmp-crun.vt5H4G.mount: Deactivated successfully. Nov 27 03:19:38 localhost podman[74513]: 2025-11-27 08:19:38.00059158 +0000 UTC m=+0.090668208 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=starting, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, tcib_managed=true, version=17.1.12, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, batch=17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, url=https://www.redhat.com) Nov 27 03:19:38 localhost podman[74514]: 2025-11-27 08:19:38.043112897 +0000 UTC m=+0.130076531 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=starting, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, vendor=Red Hat, Inc., config_id=tripleo_step4, io.openshift.expose-services=, release=1761123044, container_name=ovn_controller, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:19:38 localhost systemd[1]: tmp-crun.IJXHNY.mount: Deactivated successfully. Nov 27 03:19:38 localhost podman[74514]: 2025-11-27 08:19:38.070996529 +0000 UTC m=+0.157960213 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, release=1761123044, vendor=Red Hat, Inc., tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, version=17.1.12) Nov 27 03:19:38 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:19:38 localhost podman[74513]: 2025-11-27 08:19:38.083075055 +0000 UTC m=+0.173151663 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, build-date=2025-11-19T00:14:25Z, release=1761123044, io.buildah.version=1.41.4, io.openshift.expose-services=, url=https://www.redhat.com, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, vendor=Red Hat, Inc., tcib_managed=true, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:19:38 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:19:44 localhost snmpd[71019]: empty variable list in _query Nov 27 03:19:44 localhost snmpd[71019]: empty variable list in _query Nov 27 03:19:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:19:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:19:47 localhost podman[74637]: 2025-11-27 08:19:47.991082935 +0000 UTC m=+0.084572493 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, tcib_managed=true, io.buildah.version=1.41.4, version=17.1.12, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container) Nov 27 03:19:48 localhost podman[74638]: 2025-11-27 08:19:48.037783436 +0000 UTC m=+0.130028919 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, distribution-scope=public, vendor=Red Hat, Inc., url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, vcs-type=git) Nov 27 03:19:48 localhost podman[74637]: 2025-11-27 08:19:48.060990632 +0000 UTC m=+0.154480160 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, version=17.1.12, container_name=collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.buildah.version=1.41.4, config_id=tripleo_step3, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:19:48 localhost podman[74638]: 2025-11-27 08:19:48.075219586 +0000 UTC m=+0.167465009 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, container_name=iscsid, url=https://www.redhat.com, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:19:48 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:19:48 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:19:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:19:58 localhost podman[74675]: 2025-11-27 08:19:57.998932029 +0000 UTC m=+0.092163476 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, version=17.1.12, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, tcib_managed=true, config_id=tripleo_step1, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.expose-services=, batch=17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:19:58 localhost podman[74675]: 2025-11-27 08:19:58.187960966 +0000 UTC m=+0.281192443 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., container_name=metrics_qdr, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, architecture=x86_64, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, release=1761123044, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, vcs-type=git, distribution-scope=public, tcib_managed=true) Nov 27 03:19:58 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:20:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:20:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:20:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:20:03 localhost systemd[1]: tmp-crun.e38PwU.mount: Deactivated successfully. Nov 27 03:20:04 localhost podman[74705]: 2025-11-27 08:20:04.001340083 +0000 UTC m=+0.093226714 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, config_id=tripleo_step4, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044) Nov 27 03:20:04 localhost podman[74705]: 2025-11-27 08:20:04.040880676 +0000 UTC m=+0.132767317 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, release=1761123044, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, tcib_managed=true, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, distribution-scope=public, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:20:04 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:20:04 localhost podman[74704]: 2025-11-27 08:20:04.044239146 +0000 UTC m=+0.138288226 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, tcib_managed=true, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, container_name=ceilometer_agent_ipmi, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_step4, version=17.1.12, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:20:04 localhost podman[74706]: 2025-11-27 08:20:04.106840247 +0000 UTC m=+0.195432270 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, container_name=ceilometer_agent_compute, release=1761123044, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, vcs-type=git, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible) Nov 27 03:20:04 localhost podman[74704]: 2025-11-27 08:20:04.128038197 +0000 UTC m=+0.222087267 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, container_name=ceilometer_agent_ipmi, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-ipmi, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, io.buildah.version=1.41.4, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:20:04 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:20:04 localhost podman[74706]: 2025-11-27 08:20:04.147918711 +0000 UTC m=+0.236510724 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, config_id=tripleo_step4, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, maintainer=OpenStack TripleO Team, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, tcib_managed=true, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, release=1761123044) Nov 27 03:20:04 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:20:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:20:04 localhost podman[74776]: 2025-11-27 08:20:04.961990745 +0000 UTC m=+0.065084749 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, config_id=tripleo_step4, version=17.1.12, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., vcs-type=git, distribution-scope=public, managed_by=tripleo_ansible, container_name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:20:05 localhost podman[74776]: 2025-11-27 08:20:05.275873895 +0000 UTC m=+0.378967919 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, distribution-scope=public, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, batch=17.1_20251118.1, vcs-type=git, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., version=17.1.12, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:20:05 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:20:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:20:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:20:08 localhost systemd[1]: tmp-crun.yZoW1i.mount: Deactivated successfully. Nov 27 03:20:09 localhost podman[74799]: 2025-11-27 08:20:09.00375129 +0000 UTC m=+0.095429254 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., tcib_managed=true, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:20:09 localhost podman[74800]: 2025-11-27 08:20:09.055347316 +0000 UTC m=+0.142266172 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, batch=17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, version=17.1.12, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, release=1761123044) Nov 27 03:20:09 localhost podman[74800]: 2025-11-27 08:20:09.07821764 +0000 UTC m=+0.165136466 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., architecture=x86_64, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, build-date=2025-11-18T23:34:05Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 03:20:09 localhost podman[74799]: 2025-11-27 08:20:09.085855514 +0000 UTC m=+0.177533458 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, batch=17.1_20251118.1, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:20:09 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:20:09 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:20:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:20:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:20:18 localhost podman[74847]: 2025-11-27 08:20:18.989642173 +0000 UTC m=+0.085717994 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, tcib_managed=true, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:20:19 localhost podman[74847]: 2025-11-27 08:20:19.002883878 +0000 UTC m=+0.098959729 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, container_name=collectd, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, tcib_managed=true) Nov 27 03:20:19 localhost podman[74848]: 2025-11-27 08:20:19.037078447 +0000 UTC m=+0.129900730 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, version=17.1.12, batch=17.1_20251118.1, tcib_managed=true, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, distribution-scope=public, container_name=iscsid, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:20:19 localhost podman[74848]: 2025-11-27 08:20:19.048823533 +0000 UTC m=+0.141645776 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, release=1761123044, build-date=2025-11-18T23:44:13Z, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=iscsid, batch=17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, tcib_managed=true, name=rhosp17/openstack-iscsid) Nov 27 03:20:19 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:20:19 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:20:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:20:28 localhost podman[74889]: 2025-11-27 08:20:28.980687805 +0000 UTC m=+0.080794962 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, managed_by=tripleo_ansible, release=1761123044, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, vcs-type=git, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:20:29 localhost podman[74889]: 2025-11-27 08:20:29.164233565 +0000 UTC m=+0.264340752 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12, io.buildah.version=1.41.4, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:20:29 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:20:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:20:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:20:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:20:34 localhost systemd[1]: tmp-crun.fLdMN3.mount: Deactivated successfully. Nov 27 03:20:34 localhost podman[74918]: 2025-11-27 08:20:34.997475045 +0000 UTC m=+0.089807933 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, architecture=x86_64, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, managed_by=tripleo_ansible, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:20:35 localhost podman[74918]: 2025-11-27 08:20:35.026961277 +0000 UTC m=+0.119294175 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, architecture=x86_64, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-type=git, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, managed_by=tripleo_ansible, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, version=17.1.12, tcib_managed=true, build-date=2025-11-19T00:12:45Z, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi) Nov 27 03:20:35 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:20:35 localhost podman[74919]: 2025-11-27 08:20:35.042174386 +0000 UTC m=+0.131263997 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, batch=17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, architecture=x86_64, release=1761123044, tcib_managed=true, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, name=rhosp17/openstack-cron, vcs-type=git, com.redhat.component=openstack-cron-container, distribution-scope=public, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:20:35 localhost podman[74919]: 2025-11-27 08:20:35.07324156 +0000 UTC m=+0.162331191 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, architecture=x86_64, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, distribution-scope=public, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:20:35 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:20:35 localhost podman[74920]: 2025-11-27 08:20:35.145726447 +0000 UTC m=+0.232581208 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, container_name=ceilometer_agent_compute, architecture=x86_64, url=https://www.redhat.com, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:20:35 localhost podman[74920]: 2025-11-27 08:20:35.19609705 +0000 UTC m=+0.282951841 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, name=rhosp17/openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com) Nov 27 03:20:35 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:20:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:20:35 localhost podman[74991]: 2025-11-27 08:20:35.975561895 +0000 UTC m=+0.075794047 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044) Nov 27 03:20:36 localhost podman[74991]: 2025-11-27 08:20:36.350079914 +0000 UTC m=+0.450312066 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, io.openshift.expose-services=, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, release=1761123044, managed_by=tripleo_ansible, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:20:36 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:20:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:20:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:20:39 localhost podman[75014]: 2025-11-27 08:20:39.980678156 +0000 UTC m=+0.079698342 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, batch=17.1_20251118.1, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, io.buildah.version=1.41.4, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:20:40 localhost podman[75015]: 2025-11-27 08:20:40.032990362 +0000 UTC m=+0.129704976 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, io.buildah.version=1.41.4, tcib_managed=true, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., config_id=tripleo_step4, url=https://www.redhat.com, container_name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:20:40 localhost podman[75014]: 2025-11-27 08:20:40.047983004 +0000 UTC m=+0.147003170 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, io.buildah.version=1.41.4, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vcs-type=git, config_id=tripleo_step4) Nov 27 03:20:40 localhost podman[75015]: 2025-11-27 08:20:40.056648307 +0000 UTC m=+0.153362851 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, version=17.1.12, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public) Nov 27 03:20:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:20:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:20:41 localhost podman[75164]: 2025-11-27 08:20:41.134688001 +0000 UTC m=+0.100822199 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, version=7, io.openshift.expose-services=, distribution-scope=public, GIT_CLEAN=True, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., RELEASE=main) Nov 27 03:20:41 localhost podman[75164]: 2025-11-27 08:20:41.244020817 +0000 UTC m=+0.210155015 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, maintainer=Guillaume Abrioux , version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., RELEASE=main, release=553, CEPH_POINT_RELEASE=, architecture=x86_64, vcs-type=git, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, com.redhat.component=rhceph-container, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=) Nov 27 03:20:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:20:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:20:49 localhost systemd[1]: tmp-crun.JKgA79.mount: Deactivated successfully. Nov 27 03:20:50 localhost systemd[1]: tmp-crun.pw6DCC.mount: Deactivated successfully. Nov 27 03:20:50 localhost podman[75308]: 2025-11-27 08:20:50.051636082 +0000 UTC m=+0.142310073 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.component=openstack-iscsid-container, tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_id=tripleo_step3, vcs-type=git, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:20:50 localhost podman[75307]: 2025-11-27 08:20:50.011475613 +0000 UTC m=+0.106087939 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, com.redhat.component=openstack-collectd-container, tcib_managed=true, distribution-scope=public, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, maintainer=OpenStack TripleO Team, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd) Nov 27 03:20:50 localhost podman[75307]: 2025-11-27 08:20:50.095028417 +0000 UTC m=+0.189640723 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, distribution-scope=public, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, tcib_managed=true, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, release=1761123044, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com) Nov 27 03:20:50 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:20:50 localhost podman[75308]: 2025-11-27 08:20:50.11410312 +0000 UTC m=+0.204777101 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, architecture=x86_64, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, container_name=iscsid, release=1761123044, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, tcib_managed=true) Nov 27 03:20:50 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:20:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:20:59 localhost podman[75343]: 2025-11-27 08:20:59.981020927 +0000 UTC m=+0.077011719 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, version=17.1.12, batch=17.1_20251118.1, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., url=https://www.redhat.com, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:21:00 localhost podman[75343]: 2025-11-27 08:21:00.167935137 +0000 UTC m=+0.263925939 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, architecture=x86_64, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=tripleo_step1, io.openshift.expose-services=, batch=17.1_20251118.1) Nov 27 03:21:00 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:21:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:21:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:21:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:21:05 localhost podman[75372]: 2025-11-27 08:21:05.995604076 +0000 UTC m=+0.089030341 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64) Nov 27 03:21:06 localhost podman[75372]: 2025-11-27 08:21:06.073100558 +0000 UTC m=+0.166526803 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-type=git, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, container_name=ceilometer_agent_ipmi, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, url=https://www.redhat.com) Nov 27 03:21:06 localhost systemd[1]: tmp-crun.zvpLIH.mount: Deactivated successfully. Nov 27 03:21:06 localhost podman[75373]: 2025-11-27 08:21:06.091211204 +0000 UTC m=+0.181709181 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, architecture=x86_64, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, container_name=logrotate_crond, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vendor=Red Hat, Inc., url=https://www.redhat.com, distribution-scope=public, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, com.redhat.component=openstack-cron-container) Nov 27 03:21:06 localhost podman[75374]: 2025-11-27 08:21:06.109105415 +0000 UTC m=+0.196113858 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.4, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, batch=17.1_20251118.1, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:21:06 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:21:06 localhost podman[75374]: 2025-11-27 08:21:06.147422664 +0000 UTC m=+0.234431107 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, architecture=x86_64, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., tcib_managed=true, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, release=1761123044, maintainer=OpenStack TripleO Team, config_id=tripleo_step4) Nov 27 03:21:06 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:21:06 localhost podman[75373]: 2025-11-27 08:21:06.179226099 +0000 UTC m=+0.269724076 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, container_name=logrotate_crond, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:21:06 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:21:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:21:06 localhost podman[75444]: 2025-11-27 08:21:06.98313293 +0000 UTC m=+0.080135954 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, version=17.1.12, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute) Nov 27 03:21:07 localhost podman[75444]: 2025-11-27 08:21:07.381013176 +0000 UTC m=+0.478016140 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, name=rhosp17/openstack-nova-compute, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, version=17.1.12, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:21:07 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:21:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:21:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:21:10 localhost podman[75468]: 2025-11-27 08:21:10.987723555 +0000 UTC m=+0.081261274 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, distribution-scope=public, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_controller, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team) Nov 27 03:21:11 localhost podman[75467]: 2025-11-27 08:21:11.033637588 +0000 UTC m=+0.128754129 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12) Nov 27 03:21:11 localhost podman[75468]: 2025-11-27 08:21:11.064030465 +0000 UTC m=+0.157568174 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.openshift.expose-services=, architecture=x86_64, container_name=ovn_controller, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, url=https://www.redhat.com, version=17.1.12, tcib_managed=true, distribution-scope=public, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team) Nov 27 03:21:11 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:21:11 localhost podman[75467]: 2025-11-27 08:21:11.104553193 +0000 UTC m=+0.199669734 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, config_id=tripleo_step4, vcs-type=git, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, managed_by=tripleo_ansible, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:21:11 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:21:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:21:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:21:20 localhost podman[75513]: 2025-11-27 08:21:20.985814885 +0000 UTC m=+0.081817439 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, architecture=x86_64, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, vendor=Red Hat, Inc., release=1761123044, build-date=2025-11-18T22:51:28Z, batch=17.1_20251118.1) Nov 27 03:21:21 localhost podman[75513]: 2025-11-27 08:21:21.024409851 +0000 UTC m=+0.120412445 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, architecture=x86_64, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, container_name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, io.buildah.version=1.41.4, url=https://www.redhat.com, io.openshift.expose-services=) Nov 27 03:21:21 localhost systemd[1]: tmp-crun.H8yljO.mount: Deactivated successfully. Nov 27 03:21:21 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:21:21 localhost podman[75514]: 2025-11-27 08:21:21.04447988 +0000 UTC m=+0.137731440 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., distribution-scope=public, architecture=x86_64, release=1761123044, container_name=iscsid, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:21:21 localhost podman[75514]: 2025-11-27 08:21:21.140569822 +0000 UTC m=+0.233821442 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, distribution-scope=public, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, vcs-type=git, io.openshift.expose-services=, version=17.1.12, maintainer=OpenStack TripleO Team, url=https://www.redhat.com) Nov 27 03:21:21 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:21:24 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:21:24 localhost recover_tripleo_nova_virtqemud[75554]: 63639 Nov 27 03:21:24 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:21:24 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:21:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:21:30 localhost systemd[1]: tmp-crun.4YI9Og.mount: Deactivated successfully. Nov 27 03:21:30 localhost podman[75555]: 2025-11-27 08:21:30.98148308 +0000 UTC m=+0.076758762 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, batch=17.1_20251118.1, architecture=x86_64) Nov 27 03:21:31 localhost podman[75555]: 2025-11-27 08:21:31.175779079 +0000 UTC m=+0.271054681 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, version=17.1.12, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, container_name=metrics_qdr, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, release=1761123044, architecture=x86_64) Nov 27 03:21:31 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:21:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:21:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:21:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:21:36 localhost podman[75582]: 2025-11-27 08:21:36.97891967 +0000 UTC m=+0.072125117 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, io.buildah.version=1.41.4, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, tcib_managed=true, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:21:37 localhost podman[75582]: 2025-11-27 08:21:37.03624244 +0000 UTC m=+0.129447897 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., batch=17.1_20251118.1, version=17.1.12, tcib_managed=true, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:21:37 localhost systemd[1]: tmp-crun.qxE6ls.mount: Deactivated successfully. Nov 27 03:21:37 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:21:37 localhost podman[75583]: 2025-11-27 08:21:37.092649065 +0000 UTC m=+0.184003403 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, build-date=2025-11-18T22:49:32Z, distribution-scope=public, vendor=Red Hat, Inc., managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:21:37 localhost podman[75583]: 2025-11-27 08:21:37.126947846 +0000 UTC m=+0.218302144 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vendor=Red Hat, Inc., managed_by=tripleo_ansible, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, name=rhosp17/openstack-cron, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, release=1761123044) Nov 27 03:21:37 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:21:37 localhost podman[75584]: 2025-11-27 08:21:37.043047743 +0000 UTC m=+0.133539457 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, managed_by=tripleo_ansible, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, vendor=Red Hat, Inc., io.buildah.version=1.41.4, url=https://www.redhat.com, distribution-scope=public, maintainer=OpenStack TripleO Team) Nov 27 03:21:37 localhost podman[75584]: 2025-11-27 08:21:37.173790575 +0000 UTC m=+0.264282269 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true, config_id=tripleo_step4, distribution-scope=public, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:21:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:21:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:21:37 localhost podman[75651]: 2025-11-27 08:21:37.958621803 +0000 UTC m=+0.062849648 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, managed_by=tripleo_ansible, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, maintainer=OpenStack TripleO Team, container_name=nova_migration_target) Nov 27 03:21:38 localhost podman[75651]: 2025-11-27 08:21:38.337996182 +0000 UTC m=+0.442224087 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, version=17.1.12, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, architecture=x86_64, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z) Nov 27 03:21:38 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:21:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:21:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:21:42 localhost systemd[1]: tmp-crun.7a9Rd7.mount: Deactivated successfully. Nov 27 03:21:42 localhost podman[75675]: 2025-11-27 08:21:42.025965985 +0000 UTC m=+0.118498344 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, io.openshift.expose-services=, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_controller, distribution-scope=public, batch=17.1_20251118.1, vcs-type=git, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:21:42 localhost podman[75674]: 2025-11-27 08:21:41.986859654 +0000 UTC m=+0.086371431 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, distribution-scope=public, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, architecture=x86_64, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, url=https://www.redhat.com, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, release=1761123044, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:21:42 localhost podman[75675]: 2025-11-27 08:21:42.050104533 +0000 UTC m=+0.142636942 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, url=https://www.redhat.com, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:21:42 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:21:42 localhost podman[75674]: 2025-11-27 08:21:42.069230366 +0000 UTC m=+0.168742133 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, distribution-scope=public, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, build-date=2025-11-19T00:14:25Z, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 03:21:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:21:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:21:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:21:51 localhost podman[75800]: 2025-11-27 08:21:51.995598662 +0000 UTC m=+0.090694777 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vcs-type=git, name=rhosp17/openstack-iscsid, container_name=iscsid, distribution-scope=public, io.openshift.expose-services=, url=https://www.redhat.com, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:21:52 localhost podman[75800]: 2025-11-27 08:21:52.042750769 +0000 UTC m=+0.137846884 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, url=https://www.redhat.com, name=rhosp17/openstack-iscsid, container_name=iscsid, io.openshift.expose-services=) Nov 27 03:21:52 localhost systemd[1]: tmp-crun.PlyyiD.mount: Deactivated successfully. Nov 27 03:21:52 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:21:52 localhost podman[75799]: 2025-11-27 08:21:52.055883231 +0000 UTC m=+0.150646867 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, vcs-type=git, io.openshift.expose-services=, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, distribution-scope=public) Nov 27 03:21:52 localhost podman[75799]: 2025-11-27 08:21:52.094140139 +0000 UTC m=+0.188903735 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, build-date=2025-11-18T22:51:28Z, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, architecture=x86_64, version=17.1.12, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.openshift.expose-services=) Nov 27 03:21:52 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:22:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:22:01 localhost podman[75836]: 2025-11-27 08:22:01.985808512 +0000 UTC m=+0.085677002 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, tcib_managed=true, config_id=tripleo_step1, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, url=https://www.redhat.com, version=17.1.12, vcs-type=git, release=1761123044, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:02 localhost podman[75836]: 2025-11-27 08:22:02.195058343 +0000 UTC m=+0.294926763 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_id=tripleo_step1, distribution-scope=public) Nov 27 03:22:02 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:22:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:22:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:22:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:22:07 localhost systemd[1]: tmp-crun.GFhi39.mount: Deactivated successfully. Nov 27 03:22:07 localhost podman[75865]: 2025-11-27 08:22:07.987061734 +0000 UTC m=+0.083429962 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, io.buildah.version=1.41.4, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, version=17.1.12) Nov 27 03:22:08 localhost podman[75866]: 2025-11-27 08:22:08.042550464 +0000 UTC m=+0.134979466 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, distribution-scope=public, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, url=https://www.redhat.com, architecture=x86_64, batch=17.1_20251118.1, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:08 localhost podman[75866]: 2025-11-27 08:22:08.054841674 +0000 UTC m=+0.147270686 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, architecture=x86_64, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044, io.openshift.expose-services=, maintainer=OpenStack TripleO Team) Nov 27 03:22:08 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:22:08 localhost podman[75865]: 2025-11-27 08:22:08.070023592 +0000 UTC m=+0.166391880 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, io.buildah.version=1.41.4, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, distribution-scope=public, version=17.1.12, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:22:08 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:22:08 localhost podman[75867]: 2025-11-27 08:22:08.148076119 +0000 UTC m=+0.238578680 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, release=1761123044, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, version=17.1.12, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, container_name=ceilometer_agent_compute) Nov 27 03:22:08 localhost podman[75867]: 2025-11-27 08:22:08.178823695 +0000 UTC m=+0.269326256 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, tcib_managed=true, distribution-scope=public, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:22:08 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:22:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:22:08 localhost podman[75940]: 2025-11-27 08:22:08.97691134 +0000 UTC m=+0.077135873 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, managed_by=tripleo_ansible, architecture=x86_64, version=17.1.12, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:22:09 localhost podman[75940]: 2025-11-27 08:22:09.307069117 +0000 UTC m=+0.407293600 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, container_name=nova_migration_target, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64) Nov 27 03:22:09 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:22:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:22:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:22:12 localhost systemd[1]: tmp-crun.bUxhh9.mount: Deactivated successfully. Nov 27 03:22:12 localhost podman[75963]: 2025-11-27 08:22:12.996884979 +0000 UTC m=+0.090253025 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, tcib_managed=true, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, io.buildah.version=1.41.4, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, version=17.1.12) Nov 27 03:22:13 localhost podman[75963]: 2025-11-27 08:22:13.043876371 +0000 UTC m=+0.137244377 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public) Nov 27 03:22:13 localhost systemd[1]: tmp-crun.o4idQg.mount: Deactivated successfully. Nov 27 03:22:13 localhost podman[75964]: 2025-11-27 08:22:13.052333778 +0000 UTC m=+0.142123477 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, version=17.1.12, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, architecture=x86_64, release=1761123044, tcib_managed=true, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:22:13 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:22:13 localhost podman[75964]: 2025-11-27 08:22:13.079067717 +0000 UTC m=+0.168857476 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, build-date=2025-11-18T23:34:05Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, name=rhosp17/openstack-ovn-controller, version=17.1.12, managed_by=tripleo_ansible, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:22:13 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:22:17 localhost python3[76058]: ansible-ansible.legacy.stat Invoked with path=/etc/puppet/hieradata/config_step.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:17 localhost python3[76103]: ansible-ansible.legacy.copy Invoked with dest=/etc/puppet/hieradata/config_step.json force=True mode=0600 src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231737.1632466-114111-133523899558737/source _original_basename=tmp2ancrod8 follow=False checksum=039e0b234f00fbd1242930f0d5dc67e8b4c067fe backup=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:18 localhost python3[76133]: ansible-stat Invoked with path=/var/lib/tripleo-config/container-startup-config/step_5 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:22:20 localhost ansible-async_wrapper.py[76305]: Invoked with 280344622407 3600 /home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231740.1278763-114264-20878744120872/AnsiballZ_command.py _ Nov 27 03:22:20 localhost ansible-async_wrapper.py[76308]: Starting module and watcher Nov 27 03:22:20 localhost ansible-async_wrapper.py[76308]: Start watching 76309 (3600) Nov 27 03:22:20 localhost ansible-async_wrapper.py[76309]: Start module (76309) Nov 27 03:22:20 localhost ansible-async_wrapper.py[76305]: Return async_wrapper task started. Nov 27 03:22:21 localhost python3[76327]: ansible-ansible.legacy.async_status Invoked with jid=280344622407.76305 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:22:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:22:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:22:22 localhost podman[76350]: 2025-11-27 08:22:22.986898343 +0000 UTC m=+0.076040404 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, tcib_managed=true, architecture=x86_64, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, com.redhat.component=openstack-iscsid-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, vendor=Red Hat, Inc., batch=17.1_20251118.1, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:22:23 localhost podman[76350]: 2025-11-27 08:22:22.999958543 +0000 UTC m=+0.089100574 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, tcib_managed=true, release=1761123044, build-date=2025-11-18T23:44:13Z, container_name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, vcs-type=git, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, batch=17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, version=17.1.12, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:22:23 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:22:23 localhost podman[76349]: 2025-11-27 08:22:23.054772495 +0000 UTC m=+0.144535653 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12) Nov 27 03:22:23 localhost podman[76349]: 2025-11-27 08:22:23.094064621 +0000 UTC m=+0.183827769 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, architecture=x86_64, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, managed_by=tripleo_ansible, release=1761123044, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., tcib_managed=true, name=rhosp17/openstack-collectd, config_id=tripleo_step3, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:22:23 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:22:24 localhost puppet-user[76329]: Warning: /etc/puppet/hiera.yaml: Use of 'hiera.yaml' version 3 is deprecated. It should be converted to version 5 Nov 27 03:22:24 localhost puppet-user[76329]: (file: /etc/puppet/hiera.yaml) Nov 27 03:22:24 localhost puppet-user[76329]: Warning: Undefined variable '::deploy_config_name'; Nov 27 03:22:24 localhost puppet-user[76329]: (file & line not available) Nov 27 03:22:24 localhost puppet-user[76329]: Warning: The function 'hiera' is deprecated in favor of using 'lookup'. See https://puppet.com/docs/puppet/7.10/deprecated_language.html Nov 27 03:22:24 localhost puppet-user[76329]: (file & line not available) Nov 27 03:22:24 localhost puppet-user[76329]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/profile/base/database/mysql/client.pp, line: 89, column: 8) Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use match expressions with Stdlib::Compat::String instead. They are described at https://docs.puppet.com/puppet/latest/reference/lang_data_type.html#match-expressions. at ["/etc/puppet/modules/snmp/manifests/params.pp", 310]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:22:24 localhost puppet-user[76329]: with Stdlib::Compat::Bool. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 358]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:22:24 localhost puppet-user[76329]: with Stdlib::Compat::Array. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 367]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:22:24 localhost puppet-user[76329]: with Stdlib::Compat::String. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 382]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:22:24 localhost puppet-user[76329]: with Stdlib::Compat::Numeric. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 388]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:24 localhost puppet-user[76329]: Warning: This method is deprecated, please use the stdlib validate_legacy function, Nov 27 03:22:24 localhost puppet-user[76329]: with Pattern[]. There is further documentation for validate_legacy function in the README. at ["/etc/puppet/modules/snmp/manifests/init.pp", 393]:["/var/lib/tripleo-config/puppet_step_config.pp", 4] Nov 27 03:22:24 localhost puppet-user[76329]: (location: /etc/puppet/modules/stdlib/lib/puppet/functions/deprecation.rb:34:in `deprecation') Nov 27 03:22:25 localhost puppet-user[76329]: Warning: Unknown variable: '::deployment_type'. (file: /etc/puppet/modules/tripleo/manifests/packages.pp, line: 39, column: 69) Nov 27 03:22:25 localhost puppet-user[76329]: Notice: Compiled catalog for np0005537446.localdomain in environment production in 0.25 seconds Nov 27 03:22:25 localhost puppet-user[76329]: Notice: Applied catalog in 0.31 seconds Nov 27 03:22:25 localhost puppet-user[76329]: Application: Nov 27 03:22:25 localhost puppet-user[76329]: Initial environment: production Nov 27 03:22:25 localhost puppet-user[76329]: Converged environment: production Nov 27 03:22:25 localhost puppet-user[76329]: Run mode: user Nov 27 03:22:25 localhost puppet-user[76329]: Changes: Nov 27 03:22:25 localhost puppet-user[76329]: Events: Nov 27 03:22:25 localhost puppet-user[76329]: Resources: Nov 27 03:22:25 localhost puppet-user[76329]: Total: 19 Nov 27 03:22:25 localhost puppet-user[76329]: Time: Nov 27 03:22:25 localhost puppet-user[76329]: Schedule: 0.00 Nov 27 03:22:25 localhost puppet-user[76329]: Package: 0.00 Nov 27 03:22:25 localhost puppet-user[76329]: Exec: 0.01 Nov 27 03:22:25 localhost puppet-user[76329]: Augeas: 0.01 Nov 27 03:22:25 localhost puppet-user[76329]: File: 0.02 Nov 27 03:22:25 localhost puppet-user[76329]: Service: 0.07 Nov 27 03:22:25 localhost puppet-user[76329]: Transaction evaluation: 0.30 Nov 27 03:22:25 localhost puppet-user[76329]: Catalog application: 0.31 Nov 27 03:22:25 localhost puppet-user[76329]: Config retrieval: 0.32 Nov 27 03:22:25 localhost puppet-user[76329]: Last run: 1764231745 Nov 27 03:22:25 localhost puppet-user[76329]: Filebucket: 0.00 Nov 27 03:22:25 localhost puppet-user[76329]: Total: 0.32 Nov 27 03:22:25 localhost puppet-user[76329]: Version: Nov 27 03:22:25 localhost puppet-user[76329]: Config: 1764231744 Nov 27 03:22:25 localhost puppet-user[76329]: Puppet: 7.10.0 Nov 27 03:22:25 localhost ansible-async_wrapper.py[76309]: Module complete (76309) Nov 27 03:22:25 localhost ansible-async_wrapper.py[76308]: Done in kid B. Nov 27 03:22:31 localhost python3[76505]: ansible-ansible.legacy.async_status Invoked with jid=280344622407.76305 mode=status _async_dir=/tmp/.ansible_async Nov 27 03:22:32 localhost python3[76521]: ansible-file Invoked with path=/var/lib/container-puppet/puppetlabs state=directory setype=svirt_sandbox_file_t selevel=s0 recurse=True force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:22:32 localhost python3[76537]: ansible-stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:22:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:22:32 localhost systemd[1]: tmp-crun.TAf7oU.mount: Deactivated successfully. Nov 27 03:22:32 localhost podman[76587]: 2025-11-27 08:22:32.822632601 +0000 UTC m=+0.103144921 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, io.buildah.version=1.41.4, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}) Nov 27 03:22:32 localhost python3[76588]: ansible-ansible.legacy.stat Invoked with path=/var/lib/container-puppet/puppetlabs/facter.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:33 localhost podman[76587]: 2025-11-27 08:22:33.045979119 +0000 UTC m=+0.326491419 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., vcs-type=git, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, architecture=x86_64, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:33 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:22:33 localhost python3[76634]: ansible-ansible.legacy.file Invoked with setype=svirt_sandbox_file_t selevel=s0 dest=/var/lib/container-puppet/puppetlabs/facter.conf _original_basename=tmpubssh_e2 recurse=False state=file path=/var/lib/container-puppet/puppetlabs/facter.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None attributes=None Nov 27 03:22:33 localhost python3[76664]: ansible-file Invoked with path=/opt/puppetlabs/facter state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:34 localhost python3[76769]: ansible-ansible.posix.synchronize Invoked with src=/opt/puppetlabs/ dest=/var/lib/container-puppet/puppetlabs/ _local_rsync_path=rsync _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] ssh_connection_multiplexing=False partial=False verify_host=False mode=push dest_port=None private_key=None recursive=None links=None perms=None times=None owner=None group=None ssh_args=None link_dest=None Nov 27 03:22:35 localhost python3[76788]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:36 localhost python3[76820]: ansible-stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:22:37 localhost python3[76870]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-container-shutdown follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:37 localhost python3[76888]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-container-shutdown _original_basename=tripleo-container-shutdown recurse=False state=file path=/usr/libexec/tripleo-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:37 localhost python3[76950]: ansible-ansible.legacy.stat Invoked with path=/usr/libexec/tripleo-start-podman-container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:22:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:22:38 localhost podman[76969]: 2025-11-27 08:22:38.189835144 +0000 UTC m=+0.086915115 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-type=git, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, tcib_managed=true, distribution-scope=public, config_id=tripleo_step4, build-date=2025-11-18T22:49:32Z, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:22:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:22:38 localhost podman[76969]: 2025-11-27 08:22:38.204826727 +0000 UTC m=+0.101906708 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, name=rhosp17/openstack-cron, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, tcib_managed=true, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:38 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:22:38 localhost python3[76968]: ansible-ansible.legacy.file Invoked with mode=0700 owner=root group=root dest=/usr/libexec/tripleo-start-podman-container _original_basename=tripleo-start-podman-container recurse=False state=file path=/usr/libexec/tripleo-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:38 localhost systemd[1]: tmp-crun.bwvcD8.mount: Deactivated successfully. Nov 27 03:22:38 localhost podman[76988]: 2025-11-27 08:22:38.305597613 +0000 UTC m=+0.098111736 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, managed_by=tripleo_ansible, release=1761123044, io.buildah.version=1.41.4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, tcib_managed=true, maintainer=OpenStack TripleO Team) Nov 27 03:22:38 localhost podman[76988]: 2025-11-27 08:22:38.362949863 +0000 UTC m=+0.155463966 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, vcs-type=git, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, version=17.1.12, tcib_managed=true, url=https://www.redhat.com, managed_by=tripleo_ansible, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:22:38 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:22:38 localhost podman[76989]: 2025-11-27 08:22:38.366288254 +0000 UTC m=+0.150659898 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_id=tripleo_step4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, tcib_managed=true, architecture=x86_64) Nov 27 03:22:38 localhost podman[76989]: 2025-11-27 08:22:38.448859071 +0000 UTC m=+0.233230685 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, vendor=Red Hat, Inc., version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, tcib_managed=true, managed_by=tripleo_ansible) Nov 27 03:22:38 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:22:38 localhost python3[77104]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/tripleo-container-shutdown.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:39 localhost python3[77122]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/tripleo-container-shutdown.service _original_basename=tripleo-container-shutdown-service recurse=False state=file path=/usr/lib/systemd/system/tripleo-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:39 localhost systemd[1]: tmp-crun.ADJm23.mount: Deactivated successfully. Nov 27 03:22:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:22:39 localhost systemd[1]: tmp-crun.V67MWf.mount: Deactivated successfully. Nov 27 03:22:39 localhost podman[77185]: 2025-11-27 08:22:39.577375391 +0000 UTC m=+0.090897723 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, managed_by=tripleo_ansible, container_name=nova_migration_target, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step4, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, distribution-scope=public, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container) Nov 27 03:22:39 localhost python3[77184]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:39 localhost python3[77223]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset _original_basename=91-tripleo-container-shutdown-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-tripleo-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:39 localhost podman[77185]: 2025-11-27 08:22:39.945824446 +0000 UTC m=+0.459346718 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., batch=17.1_20251118.1, managed_by=tripleo_ansible, architecture=x86_64, release=1761123044, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team) Nov 27 03:22:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:22:40 localhost python3[77253]: ansible-systemd Invoked with name=tripleo-container-shutdown state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:22:40 localhost systemd[1]: Reloading. Nov 27 03:22:40 localhost systemd-rc-local-generator[77274]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:22:40 localhost systemd-sysv-generator[77277]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:22:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:22:41 localhost python3[77339]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system/netns-placeholder.service follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:41 localhost python3[77357]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/usr/lib/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:42 localhost python3[77419]: ansible-ansible.legacy.stat Invoked with path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Nov 27 03:22:42 localhost python3[77437]: ansible-ansible.legacy.file Invoked with mode=0644 owner=root group=root dest=/usr/lib/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/usr/lib/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:22:42 localhost python3[77467]: ansible-systemd Invoked with name=netns-placeholder state=started enabled=True daemon_reload=True daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:22:42 localhost systemd[1]: Reloading. Nov 27 03:22:43 localhost systemd-sysv-generator[77494]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:22:43 localhost systemd-rc-local-generator[77489]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:22:43 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:22:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:22:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:22:43 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:22:43 localhost recover_tripleo_nova_virtqemud[77510]: 63639 Nov 27 03:22:43 localhost systemd[1]: Starting Create netns directory... Nov 27 03:22:43 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:22:43 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:22:43 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 03:22:43 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 03:22:43 localhost systemd[1]: Finished Create netns directory. Nov 27 03:22:43 localhost podman[77505]: 2025-11-27 08:22:43.375583244 +0000 UTC m=+0.091394186 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, tcib_managed=true, config_id=tripleo_step4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:43 localhost podman[77506]: 2025-11-27 08:22:43.433740916 +0000 UTC m=+0.148744586 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., tcib_managed=true, vcs-type=git, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, config_id=tripleo_step4, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, release=1761123044, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:22:43 localhost podman[77505]: 2025-11-27 08:22:43.441129164 +0000 UTC m=+0.156940066 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, tcib_managed=true, maintainer=OpenStack TripleO Team, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, vendor=Red Hat, Inc., io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, url=https://www.redhat.com, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:22:43 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:22:43 localhost podman[77506]: 2025-11-27 08:22:43.457436262 +0000 UTC m=+0.172439972 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, release=1761123044, url=https://www.redhat.com, tcib_managed=true, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, version=17.1.12, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:22:43 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:22:43 localhost python3[77573]: ansible-container_puppet_config Invoked with update_config_hash_only=True no_archive=True check_mode=False config_vol_prefix=/var/lib/config-data debug=False net_host=True puppet_config= short_hostname= step=6 Nov 27 03:22:44 localhost systemd[1]: tmp-crun.1dUumJ.mount: Deactivated successfully. Nov 27 03:22:45 localhost python3[77694]: ansible-tripleo_container_manage Invoked with config_id=tripleo_step5 config_dir=/var/lib/tripleo-config/container-startup-config/step_5 config_patterns=*.json config_overrides={} concurrency=5 log_base_path=/var/log/containers/stdouts debug=False Nov 27 03:22:45 localhost podman[77732]: 2025-11-27 08:22:45.874025918 +0000 UTC m=+0.092674011 container create ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, distribution-scope=public, vcs-type=git, batch=17.1_20251118.1, container_name=nova_compute) Nov 27 03:22:45 localhost systemd[1]: Started libpod-conmon-ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.scope. Nov 27 03:22:45 localhost podman[77732]: 2025-11-27 08:22:45.82757946 +0000 UTC m=+0.046227563 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:22:45 localhost systemd[1]: tmp-crun.sbwcPr.mount: Deactivated successfully. Nov 27 03:22:45 localhost systemd[1]: Started libcrun container. Nov 27 03:22:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f/merged/var/log/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f/merged/var/lib/kolla/config_files/src-ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:22:45 localhost podman[77732]: 2025-11-27 08:22:45.975997877 +0000 UTC m=+0.194646010 container init ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, container_name=nova_compute, io.openshift.expose-services=, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, managed_by=tripleo_ansible, release=1761123044) Nov 27 03:22:45 localhost systemd[1]: tmp-crun.7jyhZR.mount: Deactivated successfully. Nov 27 03:22:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:22:46 localhost podman[77732]: 2025-11-27 08:22:46.01482866 +0000 UTC m=+0.233476763 container start ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, config_id=tripleo_step5, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, vcs-type=git, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:22:46 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:22:46 localhost python3[77694]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_compute --conmon-pidfile /run/nova_compute.pid --detach=True --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env LIBGUESTFS_BACKEND=direct --env TRIPLEO_CONFIG_HASH=2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a --healthcheck-command /openstack/healthcheck 5672 --ipc host --label config_id=tripleo_step5 --label container_name=nova_compute --label managed_by=tripleo_ansible --label config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_compute.log --network host --privileged=True --ulimit nofile=131072 --ulimit memlock=67108864 --user nova --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/log/containers/nova:/var/log/nova --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro --volume /var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro --volume /var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z --volume /dev:/dev --volume /lib/modules:/lib/modules:ro --volume /run:/run --volume /run/nova:/run/nova:z --volume /var/lib/iscsi:/var/lib/iscsi:z --volume /var/lib/libvirt:/var/lib/libvirt:shared --volume /sys/class/net:/sys/class/net --volume /sys/bus/pci:/sys/bus/pci --volume /boot:/boot:ro --volume /var/lib/nova:/var/lib/nova:shared registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:22:46 localhost systemd[1]: Created slice User Slice of UID 0. Nov 27 03:22:46 localhost systemd[1]: Starting User Runtime Directory /run/user/0... Nov 27 03:22:46 localhost systemd[1]: Finished User Runtime Directory /run/user/0. Nov 27 03:22:46 localhost systemd[1]: Starting User Manager for UID 0... Nov 27 03:22:46 localhost podman[77753]: 2025-11-27 08:22:46.158742504 +0000 UTC m=+0.122832250 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=starting, batch=17.1_20251118.1, url=https://www.redhat.com, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, maintainer=OpenStack TripleO Team, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, config_id=tripleo_step5, managed_by=tripleo_ansible, io.openshift.expose-services=) Nov 27 03:22:46 localhost podman[77753]: 2025-11-27 08:22:46.253954792 +0000 UTC m=+0.218044528 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, container_name=nova_compute, version=17.1.12, config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, release=1761123044, vendor=Red Hat, Inc., architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z) Nov 27 03:22:46 localhost systemd[77770]: Queued start job for default target Main User Target. Nov 27 03:22:46 localhost systemd[77770]: Created slice User Application Slice. Nov 27 03:22:46 localhost systemd[77770]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). Nov 27 03:22:46 localhost systemd[77770]: Started Daily Cleanup of User's Temporary Directories. Nov 27 03:22:46 localhost systemd[77770]: Reached target Paths. Nov 27 03:22:46 localhost systemd[77770]: Reached target Timers. Nov 27 03:22:46 localhost podman[77753]: unhealthy Nov 27 03:22:46 localhost systemd[77770]: Starting D-Bus User Message Bus Socket... Nov 27 03:22:46 localhost systemd[77770]: Starting Create User's Volatile Files and Directories... Nov 27 03:22:46 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:22:46 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:22:46 localhost systemd[77770]: Listening on D-Bus User Message Bus Socket. Nov 27 03:22:46 localhost systemd[77770]: Reached target Sockets. Nov 27 03:22:46 localhost systemd[77770]: Finished Create User's Volatile Files and Directories. Nov 27 03:22:46 localhost systemd[77770]: Reached target Basic System. Nov 27 03:22:46 localhost systemd[77770]: Reached target Main User Target. Nov 27 03:22:46 localhost systemd[77770]: Startup finished in 176ms. Nov 27 03:22:46 localhost systemd[1]: Started User Manager for UID 0. Nov 27 03:22:46 localhost systemd[1]: Started Session c10 of User root. Nov 27 03:22:46 localhost systemd[1]: session-c10.scope: Deactivated successfully. Nov 27 03:22:46 localhost podman[77865]: 2025-11-27 08:22:46.521098697 +0000 UTC m=+0.079207459 container create ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, container_name=nova_wait_for_compute_service, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, url=https://www.redhat.com, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:22:46 localhost systemd[1]: Started libpod-conmon-ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc.scope. Nov 27 03:22:46 localhost podman[77865]: 2025-11-27 08:22:46.480378773 +0000 UTC m=+0.038487625 image pull registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:22:46 localhost systemd[1]: Started libcrun container. Nov 27 03:22:46 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f864f691e2c21229cb1f3f963d7dfab6a5d7b3d5eca8c5621209e43e1748fff3/merged/container-config-scripts supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:46 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f864f691e2c21229cb1f3f963d7dfab6a5d7b3d5eca8c5621209e43e1748fff3/merged/var/log/nova supports timestamps until 2038 (0x7fffffff) Nov 27 03:22:46 localhost podman[77865]: 2025-11-27 08:22:46.604835906 +0000 UTC m=+0.162944668 container init ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, maintainer=OpenStack TripleO Team, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.openshift.expose-services=, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, batch=17.1_20251118.1, vendor=Red Hat, Inc., container_name=nova_wait_for_compute_service, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:22:46 localhost podman[77865]: 2025-11-27 08:22:46.615158933 +0000 UTC m=+0.173267715 container start ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, architecture=x86_64, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, config_id=tripleo_step5, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, maintainer=OpenStack TripleO Team, tcib_managed=true, container_name=nova_wait_for_compute_service, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4) Nov 27 03:22:46 localhost podman[77865]: 2025-11-27 08:22:46.615514273 +0000 UTC m=+0.173623045 container attach ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_wait_for_compute_service, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, batch=17.1_20251118.1, vcs-type=git, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5) Nov 27 03:22:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:22:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:22:54 localhost systemd[1]: tmp-crun.W7jxdv.mount: Deactivated successfully. Nov 27 03:22:54 localhost podman[77889]: 2025-11-27 08:22:54.010286394 +0000 UTC m=+0.099154274 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, architecture=x86_64, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, managed_by=tripleo_ansible, url=https://www.redhat.com, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, io.openshift.expose-services=, container_name=iscsid, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-type=git, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:22:54 localhost podman[77889]: 2025-11-27 08:22:54.023120648 +0000 UTC m=+0.111988508 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, tcib_managed=true, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:22:54 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:22:54 localhost podman[77888]: 2025-11-27 08:22:54.118930842 +0000 UTC m=+0.207962577 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., managed_by=tripleo_ansible, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, version=17.1.12, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:22:54 localhost podman[77888]: 2025-11-27 08:22:54.132084844 +0000 UTC m=+0.221116559 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, architecture=x86_64, tcib_managed=true, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, distribution-scope=public, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, build-date=2025-11-18T22:51:28Z) Nov 27 03:22:54 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:22:54 localhost systemd[1]: session-27.scope: Deactivated successfully. Nov 27 03:22:54 localhost systemd[1]: session-27.scope: Consumed 2.976s CPU time. Nov 27 03:22:54 localhost systemd-logind[761]: Session 27 logged out. Waiting for processes to exit. Nov 27 03:22:54 localhost systemd-logind[761]: Removed session 27. Nov 27 03:22:54 localhost systemd[1]: tmp-crun.27OQr4.mount: Deactivated successfully. Nov 27 03:22:56 localhost systemd[1]: Stopping User Manager for UID 0... Nov 27 03:22:56 localhost systemd[77770]: Activating special unit Exit the Session... Nov 27 03:22:56 localhost systemd[77770]: Stopped target Main User Target. Nov 27 03:22:56 localhost systemd[77770]: Stopped target Basic System. Nov 27 03:22:56 localhost systemd[77770]: Stopped target Paths. Nov 27 03:22:56 localhost systemd[77770]: Stopped target Sockets. Nov 27 03:22:56 localhost systemd[77770]: Stopped target Timers. Nov 27 03:22:56 localhost systemd[77770]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 03:22:56 localhost systemd[77770]: Closed D-Bus User Message Bus Socket. Nov 27 03:22:56 localhost systemd[77770]: Stopped Create User's Volatile Files and Directories. Nov 27 03:22:56 localhost systemd[77770]: Removed slice User Application Slice. Nov 27 03:22:56 localhost systemd[77770]: Reached target Shutdown. Nov 27 03:22:56 localhost systemd[77770]: Finished Exit the Session. Nov 27 03:22:56 localhost systemd[77770]: Reached target Exit the Session. Nov 27 03:22:56 localhost systemd[1]: user@0.service: Deactivated successfully. Nov 27 03:22:56 localhost systemd[1]: Stopped User Manager for UID 0. Nov 27 03:22:56 localhost systemd[1]: Stopping User Runtime Directory /run/user/0... Nov 27 03:22:56 localhost systemd[1]: run-user-0.mount: Deactivated successfully. Nov 27 03:22:56 localhost systemd[1]: user-runtime-dir@0.service: Deactivated successfully. Nov 27 03:22:56 localhost systemd[1]: Stopped User Runtime Directory /run/user/0. Nov 27 03:22:56 localhost systemd[1]: Removed slice User Slice of UID 0. Nov 27 03:23:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:23:03 localhost podman[77928]: 2025-11-27 08:23:03.968162504 +0000 UTC m=+0.069799095 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, config_id=tripleo_step1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:23:04 localhost podman[77928]: 2025-11-27 08:23:04.184421152 +0000 UTC m=+0.286057713 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, release=1761123044, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, name=rhosp17/openstack-qdrouterd, version=17.1.12, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, container_name=metrics_qdr, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true) Nov 27 03:23:04 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:23:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:23:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:23:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:23:08 localhost podman[77957]: 2025-11-27 08:23:08.998803458 +0000 UTC m=+0.090875852 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, distribution-scope=public, url=https://www.redhat.com, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc.) Nov 27 03:23:09 localhost podman[77957]: 2025-11-27 08:23:09.051959495 +0000 UTC m=+0.144031849 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, container_name=ceilometer_agent_ipmi, config_id=tripleo_step4, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:23:09 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:23:09 localhost podman[77958]: 2025-11-27 08:23:09.065709715 +0000 UTC m=+0.154217093 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, build-date=2025-11-18T22:49:32Z, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, architecture=x86_64, distribution-scope=public, release=1761123044, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=logrotate_crond, io.openshift.expose-services=) Nov 27 03:23:09 localhost podman[77959]: 2025-11-27 08:23:09.111500115 +0000 UTC m=+0.198565784 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, vcs-type=git, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., architecture=x86_64, config_id=tripleo_step4, batch=17.1_20251118.1, tcib_managed=true, version=17.1.12, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:23:09 localhost podman[77958]: 2025-11-27 08:23:09.128684547 +0000 UTC m=+0.217191965 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, io.openshift.expose-services=, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, tcib_managed=true, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:23:09 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:23:09 localhost podman[77959]: 2025-11-27 08:23:09.168718712 +0000 UTC m=+0.255784371 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, release=1761123044, tcib_managed=true, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12) Nov 27 03:23:09 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:23:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:23:10 localhost podman[78031]: 2025-11-27 08:23:10.992883275 +0000 UTC m=+0.082001993 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, distribution-scope=public, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, tcib_managed=true, architecture=x86_64, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, container_name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:23:11 localhost podman[78031]: 2025-11-27 08:23:11.369886801 +0000 UTC m=+0.459005589 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, distribution-scope=public, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git) Nov 27 03:23:11 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:23:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:23:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:23:13 localhost systemd[1]: tmp-crun.p9kp6P.mount: Deactivated successfully. Nov 27 03:23:13 localhost podman[78055]: 2025-11-27 08:23:13.983616411 +0000 UTC m=+0.082821106 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, distribution-scope=public, vcs-type=git, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, tcib_managed=true, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4) Nov 27 03:23:14 localhost systemd[1]: tmp-crun.nncnt5.mount: Deactivated successfully. Nov 27 03:23:14 localhost podman[78056]: 2025-11-27 08:23:14.04687488 +0000 UTC m=+0.145430197 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, version=17.1.12, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, managed_by=tripleo_ansible, distribution-scope=public, release=1761123044, build-date=2025-11-18T23:34:05Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, url=https://www.redhat.com, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 03:23:14 localhost podman[78055]: 2025-11-27 08:23:14.058005979 +0000 UTC m=+0.157210664 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, version=17.1.12, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, architecture=x86_64, release=1761123044, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4) Nov 27 03:23:14 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:23:14 localhost podman[78056]: 2025-11-27 08:23:14.099659147 +0000 UTC m=+0.198214464 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, build-date=2025-11-18T23:34:05Z, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, managed_by=tripleo_ansible, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:23:14 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:23:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:23:16 localhost podman[78106]: 2025-11-27 08:23:16.997085836 +0000 UTC m=+0.086987157 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=starting, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, architecture=x86_64, vcs-type=git, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.expose-services=, config_id=tripleo_step5, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, batch=17.1_20251118.1) Nov 27 03:23:17 localhost podman[78106]: 2025-11-27 08:23:17.053504732 +0000 UTC m=+0.143405983 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, config_id=tripleo_step5, release=1761123044, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-type=git, distribution-scope=public, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, url=https://www.redhat.com) Nov 27 03:23:17 localhost podman[78106]: unhealthy Nov 27 03:23:17 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:23:17 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:23:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:23:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:23:24 localhost podman[78129]: 2025-11-27 08:23:24.993048113 +0000 UTC m=+0.085047406 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-iscsid-container, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, release=1761123044, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:23:25 localhost systemd[1]: tmp-crun.iW90q5.mount: Deactivated successfully. Nov 27 03:23:25 localhost podman[78128]: 2025-11-27 08:23:25.060609248 +0000 UTC m=+0.151645284 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, release=1761123044, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, config_id=tripleo_step3, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, architecture=x86_64, container_name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:23:25 localhost podman[78128]: 2025-11-27 08:23:25.06811699 +0000 UTC m=+0.159153086 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vcs-type=git, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, vendor=Red Hat, Inc., architecture=x86_64, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, config_id=tripleo_step3, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, maintainer=OpenStack TripleO Team) Nov 27 03:23:25 localhost podman[78129]: 2025-11-27 08:23:25.078555309 +0000 UTC m=+0.170554642 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., release=1761123044, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, url=https://www.redhat.com, distribution-scope=public, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:23:25 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:23:25 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:23:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:23:34 localhost systemd[1]: tmp-crun.f8m7iz.mount: Deactivated successfully. Nov 27 03:23:35 localhost podman[78169]: 2025-11-27 08:23:34.997920235 +0000 UTC m=+0.094183811 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, version=17.1.12, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, batch=17.1_20251118.1, architecture=x86_64, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:23:35 localhost podman[78169]: 2025-11-27 08:23:35.216780683 +0000 UTC m=+0.313044249 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, tcib_managed=true, version=17.1.12, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, batch=17.1_20251118.1, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=) Nov 27 03:23:35 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:23:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:23:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:23:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:23:39 localhost systemd[1]: tmp-crun.o8fzMw.mount: Deactivated successfully. Nov 27 03:23:39 localhost podman[78200]: 2025-11-27 08:23:39.996031156 +0000 UTC m=+0.081427258 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, version=17.1.12, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, config_id=tripleo_step4, container_name=logrotate_crond, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, tcib_managed=true, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.buildah.version=1.41.4, architecture=x86_64, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron) Nov 27 03:23:40 localhost podman[78200]: 2025-11-27 08:23:40.03789732 +0000 UTC m=+0.123293432 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step4, vcs-type=git, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, architecture=x86_64, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond) Nov 27 03:23:40 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:23:40 localhost podman[78201]: 2025-11-27 08:23:40.013752142 +0000 UTC m=+0.092105985 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, release=1761123044, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z) Nov 27 03:23:40 localhost podman[78201]: 2025-11-27 08:23:40.092484997 +0000 UTC m=+0.170838770 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, release=1761123044, version=17.1.12, io.buildah.version=1.41.4, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, architecture=x86_64, batch=17.1_20251118.1, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:23:40 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:23:40 localhost podman[78199]: 2025-11-27 08:23:40.041895728 +0000 UTC m=+0.128040460 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, build-date=2025-11-19T00:12:45Z, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:23:40 localhost podman[78199]: 2025-11-27 08:23:40.174339405 +0000 UTC m=+0.260484167 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_id=tripleo_step4, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, batch=17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, architecture=x86_64, build-date=2025-11-19T00:12:45Z, version=17.1.12, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:23:40 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:23:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:23:41 localhost podman[78271]: 2025-11-27 08:23:41.977809283 +0000 UTC m=+0.076485006 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, vcs-type=git, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, managed_by=tripleo_ansible, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, release=1761123044, batch=17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:23:42 localhost podman[78271]: 2025-11-27 08:23:42.347955514 +0000 UTC m=+0.446631237 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, batch=17.1_20251118.1, architecture=x86_64, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git) Nov 27 03:23:42 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:23:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:23:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:23:44 localhost systemd[1]: tmp-crun.80myue.mount: Deactivated successfully. Nov 27 03:23:44 localhost podman[78295]: 2025-11-27 08:23:44.995959964 +0000 UTC m=+0.091105538 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, tcib_managed=true, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., distribution-scope=public, version=17.1.12, io.openshift.expose-services=, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:23:45 localhost podman[78296]: 2025-11-27 08:23:45.0449589 +0000 UTC m=+0.136012164 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, architecture=x86_64, release=1761123044, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:23:45 localhost podman[78295]: 2025-11-27 08:23:45.064242937 +0000 UTC m=+0.159388511 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, distribution-scope=public, version=17.1.12, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, release=1761123044) Nov 27 03:23:45 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:23:45 localhost podman[78296]: 2025-11-27 08:23:45.098833647 +0000 UTC m=+0.189886901 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, container_name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, distribution-scope=public, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., vcs-type=git, io.buildah.version=1.41.4, batch=17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 03:23:45 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:23:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:23:48 localhost podman[78404]: 2025-11-27 08:23:48.001316543 +0000 UTC m=+0.088501869 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=starting, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-type=git, config_id=tripleo_step5, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, name=rhosp17/openstack-nova-compute, architecture=x86_64, managed_by=tripleo_ansible) Nov 27 03:23:48 localhost podman[78404]: 2025-11-27 08:23:48.063984915 +0000 UTC m=+0.151170271 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., tcib_managed=true, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, architecture=x86_64, build-date=2025-11-19T00:36:58Z, vcs-type=git, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, url=https://www.redhat.com) Nov 27 03:23:48 localhost podman[78404]: unhealthy Nov 27 03:23:48 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:23:48 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:23:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:23:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:23:55 localhost systemd[1]: tmp-crun.lskr1Z.mount: Deactivated successfully. Nov 27 03:23:56 localhost podman[78441]: 2025-11-27 08:23:56.005763347 +0000 UTC m=+0.100641064 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, name=rhosp17/openstack-iscsid, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, io.openshift.expose-services=, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:23:56 localhost podman[78441]: 2025-11-27 08:23:56.04682456 +0000 UTC m=+0.141702177 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, managed_by=tripleo_ansible, batch=17.1_20251118.1, distribution-scope=public, version=17.1.12, name=rhosp17/openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, release=1761123044, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, container_name=iscsid, vcs-type=git, config_id=tripleo_step3, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:23:56 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:23:56 localhost podman[78440]: 2025-11-27 08:23:56.049601965 +0000 UTC m=+0.145943451 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, url=https://www.redhat.com, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, version=17.1.12, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:23:56 localhost podman[78440]: 2025-11-27 08:23:56.137957157 +0000 UTC m=+0.234298663 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, tcib_managed=true, distribution-scope=public, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:23:56 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:24:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:24:05 localhost podman[78479]: 2025-11-27 08:24:05.987058538 +0000 UTC m=+0.087849951 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-type=git, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, release=1761123044, build-date=2025-11-18T22:49:46Z, url=https://www.redhat.com) Nov 27 03:24:06 localhost podman[78479]: 2025-11-27 08:24:06.169831036 +0000 UTC m=+0.270622509 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vendor=Red Hat, Inc., io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-qdrouterd, vcs-type=git, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, release=1761123044, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:24:06 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:24:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:24:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:24:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:24:11 localhost systemd[1]: tmp-crun.mC3JP0.mount: Deactivated successfully. Nov 27 03:24:11 localhost podman[78510]: 2025-11-27 08:24:11.012509573 +0000 UTC m=+0.101895711 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, managed_by=tripleo_ansible, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, vcs-type=git, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:24:11 localhost podman[78510]: 2025-11-27 08:24:11.073988466 +0000 UTC m=+0.163374664 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step4, container_name=ceilometer_agent_compute, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:11:48Z, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, vendor=Red Hat, Inc., vcs-type=git) Nov 27 03:24:11 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:24:11 localhost systemd[1]: tmp-crun.3fYJN6.mount: Deactivated successfully. Nov 27 03:24:11 localhost podman[78509]: 2025-11-27 08:24:11.156740571 +0000 UTC m=+0.247582179 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, com.redhat.component=openstack-cron-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., tcib_managed=true, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, container_name=logrotate_crond, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, batch=17.1_20251118.1, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:24:11 localhost podman[78509]: 2025-11-27 08:24:11.190240402 +0000 UTC m=+0.281082010 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, version=17.1.12, container_name=logrotate_crond, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, tcib_managed=true, io.openshift.expose-services=, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com) Nov 27 03:24:11 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:24:11 localhost podman[78508]: 2025-11-27 08:24:11.194170797 +0000 UTC m=+0.288940570 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64) Nov 27 03:24:11 localhost podman[78508]: 2025-11-27 08:24:11.27421224 +0000 UTC m=+0.368982023 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12) Nov 27 03:24:11 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:24:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:24:12 localhost podman[78580]: 2025-11-27 08:24:12.990745676 +0000 UTC m=+0.081906793 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, batch=17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, tcib_managed=true) Nov 27 03:24:13 localhost podman[78580]: 2025-11-27 08:24:13.435166736 +0000 UTC m=+0.526327903 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_id=tripleo_step4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, vcs-type=git, io.openshift.expose-services=, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, release=1761123044, container_name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64) Nov 27 03:24:13 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:24:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:24:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:24:15 localhost systemd[1]: tmp-crun.wzqiMw.mount: Deactivated successfully. Nov 27 03:24:15 localhost podman[78604]: 2025-11-27 08:24:15.980749646 +0000 UTC m=+0.080885816 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, vcs-type=git, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, io.buildah.version=1.41.4, version=17.1.12, container_name=ovn_controller, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team) Nov 27 03:24:16 localhost podman[78603]: 2025-11-27 08:24:16.032779455 +0000 UTC m=+0.131222350 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, release=1761123044, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, config_id=tripleo_step4, tcib_managed=true, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 03:24:16 localhost podman[78604]: 2025-11-27 08:24:16.055206488 +0000 UTC m=+0.155342618 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, distribution-scope=public, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, architecture=x86_64, release=1761123044, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4) Nov 27 03:24:16 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:24:16 localhost podman[78603]: 2025-11-27 08:24:16.083138309 +0000 UTC m=+0.181581224 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, tcib_managed=true, io.openshift.expose-services=, config_id=tripleo_step4, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, io.buildah.version=1.41.4, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:24:16 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:24:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:24:18 localhost podman[78647]: 2025-11-27 08:24:18.969343827 +0000 UTC m=+0.071102972 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, architecture=x86_64, release=1761123044, container_name=nova_compute, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:24:19 localhost podman[78647]: 2025-11-27 08:24:19.034020887 +0000 UTC m=+0.135780082 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, tcib_managed=true, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, io.openshift.expose-services=, vendor=Red Hat, Inc.) Nov 27 03:24:19 localhost podman[78647]: unhealthy Nov 27 03:24:19 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:24:19 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:24:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:24:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:24:26 localhost podman[78670]: 2025-11-27 08:24:26.970159213 +0000 UTC m=+0.061726851 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, build-date=2025-11-18T23:44:13Z, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, vcs-type=git, release=1761123044, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid) Nov 27 03:24:26 localhost podman[78670]: 2025-11-27 08:24:26.980944833 +0000 UTC m=+0.072512471 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.component=openstack-iscsid-container, distribution-scope=public, container_name=iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, release=1761123044, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:24:26 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:24:27 localhost podman[78669]: 2025-11-27 08:24:27.028378379 +0000 UTC m=+0.120888622 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, distribution-scope=public, build-date=2025-11-18T22:51:28Z) Nov 27 03:24:27 localhost podman[78669]: 2025-11-27 08:24:27.036724723 +0000 UTC m=+0.129234946 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-18T22:51:28Z, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:24:27 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:24:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:24:36 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:24:36 localhost recover_tripleo_nova_virtqemud[78713]: 63639 Nov 27 03:24:36 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:24:36 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:24:37 localhost podman[78706]: 2025-11-27 08:24:37.014335635 +0000 UTC m=+0.105600551 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, distribution-scope=public, release=1761123044, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:24:37 localhost podman[78706]: 2025-11-27 08:24:37.219863791 +0000 UTC m=+0.311128767 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, tcib_managed=true, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, distribution-scope=public, release=1761123044, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vcs-type=git) Nov 27 03:24:37 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:24:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:24:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:24:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:24:42 localhost podman[78740]: 2025-11-27 08:24:41.999634985 +0000 UTC m=+0.086527967 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:24:42 localhost podman[78740]: 2025-11-27 08:24:42.056654239 +0000 UTC m=+0.143547241 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step4, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:24:42 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:24:42 localhost podman[78738]: 2025-11-27 08:24:42.107723012 +0000 UTC m=+0.200285926 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, tcib_managed=true, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:24:42 localhost podman[78739]: 2025-11-27 08:24:42.0619096 +0000 UTC m=+0.149359988 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, architecture=x86_64, tcib_managed=true, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, container_name=logrotate_crond, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:24:42 localhost podman[78738]: 2025-11-27 08:24:42.138851758 +0000 UTC m=+0.231414672 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:24:42 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:24:42 localhost podman[78739]: 2025-11-27 08:24:42.195294806 +0000 UTC m=+0.282745134 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, distribution-scope=public, vendor=Red Hat, Inc., url=https://www.redhat.com, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, release=1761123044, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:24:42 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:24:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:24:43 localhost podman[78809]: 2025-11-27 08:24:43.986679696 +0000 UTC m=+0.082056068 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-type=git, vendor=Red Hat, Inc., architecture=x86_64, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:24:44 localhost podman[78809]: 2025-11-27 08:24:44.369812719 +0000 UTC m=+0.465189061 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, distribution-scope=public, maintainer=OpenStack TripleO Team, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_migration_target) Nov 27 03:24:44 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:24:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:24:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:24:46 localhost podman[78831]: 2025-11-27 08:24:46.974873705 +0000 UTC m=+0.070555958 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, tcib_managed=true, managed_by=tripleo_ansible, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, architecture=x86_64, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4) Nov 27 03:24:47 localhost systemd[1]: tmp-crun.GurZ9K.mount: Deactivated successfully. Nov 27 03:24:47 localhost podman[78832]: 2025-11-27 08:24:47.0397285 +0000 UTC m=+0.132411042 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, url=https://www.redhat.com, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, distribution-scope=public, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:24:47 localhost podman[78831]: 2025-11-27 08:24:47.046831721 +0000 UTC m=+0.142513954 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, io.openshift.expose-services=, container_name=ovn_metadata_agent, batch=17.1_20251118.1, tcib_managed=true, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:24:47 localhost podman[78832]: 2025-11-27 08:24:47.069082609 +0000 UTC m=+0.161765101 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.4, release=1761123044, version=17.1.12, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:24:47 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:24:47 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:24:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:24:49 localhost podman[78951]: 2025-11-27 08:24:49.940958441 +0000 UTC m=+0.081599795 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, vcs-type=git, distribution-scope=public, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, config_id=tripleo_step5, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, tcib_managed=true, architecture=x86_64, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:24:50 localhost podman[78951]: 2025-11-27 08:24:50.004979563 +0000 UTC m=+0.145620867 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, tcib_managed=true, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, container_name=nova_compute, architecture=x86_64, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, distribution-scope=public, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container) Nov 27 03:24:50 localhost podman[78951]: unhealthy Nov 27 03:24:50 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:24:50 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:24:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:24:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:24:57 localhost podman[78976]: 2025-11-27 08:24:57.96275655 +0000 UTC m=+0.055524984 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, batch=17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, release=1761123044, distribution-scope=public, vendor=Red Hat, Inc., architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible, version=17.1.12, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:24:57 localhost podman[78976]: 2025-11-27 08:24:57.997864594 +0000 UTC m=+0.090633038 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.openshift.expose-services=, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, name=rhosp17/openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, distribution-scope=public, vendor=Red Hat, Inc., container_name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:24:58 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:24:58 localhost podman[78975]: 2025-11-27 08:24:58.087909515 +0000 UTC m=+0.185706905 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, vcs-type=git, config_id=tripleo_step3, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, version=17.1.12, distribution-scope=public, name=rhosp17/openstack-collectd, tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd) Nov 27 03:24:58 localhost podman[78975]: 2025-11-27 08:24:58.101800048 +0000 UTC m=+0.199597468 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, version=17.1.12, build-date=2025-11-18T22:51:28Z, tcib_managed=true, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:24:58 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:25:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:25:07 localhost podman[79015]: 2025-11-27 08:25:07.986163761 +0000 UTC m=+0.081480562 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, release=1761123044, managed_by=tripleo_ansible, tcib_managed=true, container_name=metrics_qdr, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, version=17.1.12) Nov 27 03:25:08 localhost podman[79015]: 2025-11-27 08:25:08.169325175 +0000 UTC m=+0.264641976 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, container_name=metrics_qdr, distribution-scope=public, managed_by=tripleo_ansible, version=17.1.12, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 03:25:08 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:25:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:25:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:25:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:25:13 localhost systemd[1]: tmp-crun.WXMs2W.mount: Deactivated successfully. Nov 27 03:25:13 localhost podman[79045]: 2025-11-27 08:25:13.017381276 +0000 UTC m=+0.110610055 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, vcs-type=git, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:25:13 localhost podman[79045]: 2025-11-27 08:25:13.100971734 +0000 UTC m=+0.194200543 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-type=git, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, architecture=x86_64, release=1761123044, vendor=Red Hat, Inc., distribution-scope=public, maintainer=OpenStack TripleO Team) Nov 27 03:25:13 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:25:13 localhost podman[79047]: 2025-11-27 08:25:13.066327112 +0000 UTC m=+0.153266322 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, tcib_managed=true, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, url=https://www.redhat.com, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:25:13 localhost podman[79047]: 2025-11-27 08:25:13.145956514 +0000 UTC m=+0.232895714 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-compute, vcs-type=git, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, architecture=x86_64, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1) Nov 27 03:25:13 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:25:13 localhost podman[79046]: 2025-11-27 08:25:13.192372951 +0000 UTC m=+0.284352166 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, version=17.1.12, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, release=1761123044, io.openshift.expose-services=, container_name=logrotate_crond, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, tcib_managed=true, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, architecture=x86_64) Nov 27 03:25:13 localhost podman[79046]: 2025-11-27 08:25:13.229947532 +0000 UTC m=+0.321926747 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, config_id=tripleo_step4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team) Nov 27 03:25:13 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:25:13 localhost systemd[1]: tmp-crun.FrGGFr.mount: Deactivated successfully. Nov 27 03:25:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:25:14 localhost podman[79118]: 2025-11-27 08:25:14.960454243 +0000 UTC m=+0.063524688 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, io.buildah.version=1.41.4, url=https://www.redhat.com, version=17.1.12, distribution-scope=public, release=1761123044, vcs-type=git, vendor=Red Hat, Inc., architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=) Nov 27 03:25:15 localhost podman[79118]: 2025-11-27 08:25:15.31550536 +0000 UTC m=+0.418575785 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, tcib_managed=true, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:25:15 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:25:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:25:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:25:17 localhost systemd[1]: tmp-crun.45tfNQ.mount: Deactivated successfully. Nov 27 03:25:17 localhost podman[79144]: 2025-11-27 08:25:17.995278687 +0000 UTC m=+0.087423782 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-type=git, architecture=x86_64, config_id=tripleo_step4, batch=17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, container_name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:25:18 localhost podman[79144]: 2025-11-27 08:25:18.023923487 +0000 UTC m=+0.116068632 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, vcs-type=git, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:25:18 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:25:18 localhost systemd[1]: tmp-crun.Huyagj.mount: Deactivated successfully. Nov 27 03:25:18 localhost podman[79143]: 2025-11-27 08:25:18.056143263 +0000 UTC m=+0.152211464 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, build-date=2025-11-19T00:14:25Z, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, release=1761123044, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, config_id=tripleo_step4, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:25:18 localhost podman[79143]: 2025-11-27 08:25:18.105989974 +0000 UTC m=+0.202058165 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, vcs-type=git, vendor=Red Hat, Inc., tcib_managed=true, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:25:18 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:25:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:25:20 localhost podman[79191]: 2025-11-27 08:25:20.981914735 +0000 UTC m=+0.082920791 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, vcs-type=git, container_name=nova_compute, batch=17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, release=1761123044, config_id=tripleo_step5, version=17.1.12, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:25:21 localhost podman[79191]: 2025-11-27 08:25:21.044901218 +0000 UTC m=+0.145907274 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, version=17.1.12, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-type=git, batch=17.1_20251118.1, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, release=1761123044, architecture=x86_64, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:25:21 localhost podman[79191]: unhealthy Nov 27 03:25:21 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:25:21 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 03:25:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:25:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:25:28 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:25:28 localhost recover_tripleo_nova_virtqemud[79223]: 63639 Nov 27 03:25:28 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:25:28 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:25:28 localhost podman[79212]: 2025-11-27 08:25:28.994442704 +0000 UTC m=+0.088894162 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, container_name=collectd, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, architecture=x86_64, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, tcib_managed=true, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-collectd-container, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:25:29 localhost podman[79212]: 2025-11-27 08:25:29.003811225 +0000 UTC m=+0.098262663 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, version=17.1.12, distribution-scope=public, config_id=tripleo_step3, architecture=x86_64, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, url=https://www.redhat.com, io.openshift.expose-services=, batch=17.1_20251118.1, container_name=collectd, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc.) Nov 27 03:25:29 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:25:29 localhost podman[79213]: 2025-11-27 08:25:29.088232026 +0000 UTC m=+0.179229150 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, url=https://www.redhat.com, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:25:29 localhost podman[79213]: 2025-11-27 08:25:29.096916919 +0000 UTC m=+0.187913983 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., url=https://www.redhat.com, config_id=tripleo_step3, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, distribution-scope=public, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, name=rhosp17/openstack-iscsid, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:25:29 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:25:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:25:38 localhost podman[79253]: 2025-11-27 08:25:38.965815516 +0000 UTC m=+0.062279566 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:25:39 localhost podman[79253]: 2025-11-27 08:25:39.161058576 +0000 UTC m=+0.257522666 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:25:39 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:25:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:25:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:25:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:25:44 localhost podman[79284]: 2025-11-27 08:25:44.002446768 +0000 UTC m=+0.096786793 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, version=17.1.12, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, architecture=x86_64, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:25:44 localhost systemd[1]: tmp-crun.5PM14Y.mount: Deactivated successfully. Nov 27 03:25:44 localhost podman[79286]: 2025-11-27 08:25:44.051590579 +0000 UTC m=+0.137910819 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.openshift.expose-services=, release=1761123044, distribution-scope=public, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, container_name=ceilometer_agent_compute) Nov 27 03:25:44 localhost podman[79286]: 2025-11-27 08:25:44.10219709 +0000 UTC m=+0.188517330 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, build-date=2025-11-19T00:11:48Z, tcib_managed=true, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:25:44 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:25:44 localhost podman[79284]: 2025-11-27 08:25:44.131159389 +0000 UTC m=+0.225499414 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_ipmi, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, io.openshift.expose-services=, managed_by=tripleo_ansible, tcib_managed=true, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 03:25:44 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:25:44 localhost podman[79285]: 2025-11-27 08:25:44.106693871 +0000 UTC m=+0.192760884 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, maintainer=OpenStack TripleO Team, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, container_name=logrotate_crond) Nov 27 03:25:44 localhost podman[79285]: 2025-11-27 08:25:44.186382004 +0000 UTC m=+0.272449037 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.expose-services=, url=https://www.redhat.com, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z) Nov 27 03:25:44 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:25:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:25:45 localhost podman[79356]: 2025-11-27 08:25:45.98774677 +0000 UTC m=+0.083140437 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, version=17.1.12, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, release=1761123044, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, batch=17.1_20251118.1, vcs-type=git, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public) Nov 27 03:25:46 localhost podman[79356]: 2025-11-27 08:25:46.387836449 +0000 UTC m=+0.483230126 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step4, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, url=https://www.redhat.com, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.expose-services=, distribution-scope=public, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:25:46 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:25:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:25:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:25:48 localhost podman[79382]: 2025-11-27 08:25:48.994659693 +0000 UTC m=+0.086877377 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, release=1761123044, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:25:49 localhost podman[79382]: 2025-11-27 08:25:49.04176266 +0000 UTC m=+0.133980324 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z) Nov 27 03:25:49 localhost podman[79383]: 2025-11-27 08:25:49.053055704 +0000 UTC m=+0.141125796 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, tcib_managed=true, config_id=tripleo_step4, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, release=1761123044, architecture=x86_64, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:25:49 localhost podman[79383]: 2025-11-27 08:25:49.081756456 +0000 UTC m=+0.169826508 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, distribution-scope=public, name=rhosp17/openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., vcs-type=git, io.buildah.version=1.41.4, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:25:49 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:25:49 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:25:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:25:51 localhost podman[79573]: 2025-11-27 08:25:51.591691715 +0000 UTC m=+0.083510626 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, container_name=nova_compute, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, architecture=x86_64, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, com.redhat.component=openstack-nova-compute-container, vcs-type=git, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., tcib_managed=true, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com) Nov 27 03:25:51 localhost podman[79573]: 2025-11-27 08:25:51.65098699 +0000 UTC m=+0.142805901 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-type=git, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, container_name=nova_compute, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, tcib_managed=true, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:25:51 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:25:59 localhost systemd[1]: libpod-ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc.scope: Deactivated successfully. Nov 27 03:25:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:25:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:25:59 localhost podman[79621]: 2025-11-27 08:25:59.248591734 +0000 UTC m=+0.064017973 container died ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, distribution-scope=public, io.buildah.version=1.41.4, container_name=nova_wait_for_compute_service, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, name=rhosp17/openstack-nova-compute, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git) Nov 27 03:25:59 localhost systemd[1]: tmp-crun.LNgL9j.mount: Deactivated successfully. Nov 27 03:25:59 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc-userdata-shm.mount: Deactivated successfully. Nov 27 03:25:59 localhost systemd[1]: var-lib-containers-storage-overlay-f864f691e2c21229cb1f3f963d7dfab6a5d7b3d5eca8c5621209e43e1748fff3-merged.mount: Deactivated successfully. Nov 27 03:25:59 localhost podman[79621]: 2025-11-27 08:25:59.283384349 +0000 UTC m=+0.098810508 container cleanup ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_wait_for_compute_service, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, distribution-scope=public, io.openshift.expose-services=, release=1761123044, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, tcib_managed=true, config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']}, architecture=x86_64, batch=17.1_20251118.1, container_name=nova_wait_for_compute_service, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:25:59 localhost systemd[1]: libpod-conmon-ab70cfe1e397be49fa0ed48e100adf279f470aae2df553822c46f38e5c2299cc.scope: Deactivated successfully. Nov 27 03:25:59 localhost python3[77694]: ansible-tripleo_container_manage PODMAN-CONTAINER-DEBUG: podman run --name nova_wait_for_compute_service --conmon-pidfile /run/nova_wait_for_compute_service.pid --detach=False --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env __OS_DEBUG=true --env TRIPLEO_CONFIG_HASH=4018231ef757a0bfd34181dc6dc2933a --label config_id=tripleo_step5 --label container_name=nova_wait_for_compute_service --label managed_by=tripleo_ansible --label config_data={'detach': False, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', '__OS_DEBUG': 'true', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'start_order': 4, 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/log/containers/nova:/var/log/nova', '/var/lib/container-config-scripts:/container-config-scripts']} --log-driver k8s-file --log-opt path=/var/log/containers/stdouts/nova_wait_for_compute_service.log --network host --user nova --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /etc/puppet:/etc/puppet:ro --volume /var/lib/kolla/config_files/nova_compute_wait_for_compute_service.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro --volume /var/log/containers/nova:/var/log/nova --volume /var/lib/container-config-scripts:/container-config-scripts registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1 Nov 27 03:25:59 localhost podman[79628]: 2025-11-27 08:25:59.337613678 +0000 UTC m=+0.138916857 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, container_name=iscsid, tcib_managed=true, com.redhat.component=openstack-iscsid-container, distribution-scope=public, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=) Nov 27 03:25:59 localhost podman[79628]: 2025-11-27 08:25:59.351958174 +0000 UTC m=+0.153261393 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.openshift.expose-services=, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.buildah.version=1.41.4, container_name=iscsid, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:25:59 localhost podman[79623]: 2025-11-27 08:25:59.347834322 +0000 UTC m=+0.150685053 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, vcs-type=git, batch=17.1_20251118.1, release=1761123044, container_name=collectd, managed_by=tripleo_ansible, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, distribution-scope=public, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:25:59 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:25:59 localhost podman[79623]: 2025-11-27 08:25:59.438017177 +0000 UTC m=+0.240867918 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-18T22:51:28Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, container_name=collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public) Nov 27 03:25:59 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:26:00 localhost python3[79713]: ansible-file Invoked with path=/etc/systemd/system/tripleo_nova_compute.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:26:00 localhost systemd[1]: tmp-crun.JTrp5m.mount: Deactivated successfully. Nov 27 03:26:00 localhost python3[79729]: ansible-stat Invoked with path=/etc/systemd/system/tripleo_nova_compute_healthcheck.timer follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 27 03:26:00 localhost python3[79790]: ansible-copy Invoked with src=/home/tripleo-admin/.ansible/tmp/ansible-tmp-1764231960.405912-118889-17552066830882/source dest=/etc/systemd/system/tripleo_nova_compute.service mode=0644 owner=root group=root backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:26:01 localhost python3[79806]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 03:26:01 localhost systemd[1]: Reloading. Nov 27 03:26:01 localhost systemd-sysv-generator[79835]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:26:01 localhost systemd-rc-local-generator[79829]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:26:01 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:26:02 localhost python3[79858]: ansible-systemd Invoked with state=restarted name=tripleo_nova_compute.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 03:26:02 localhost systemd[1]: Reloading. Nov 27 03:26:02 localhost systemd-rc-local-generator[79886]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:26:02 localhost systemd-sysv-generator[79891]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:26:02 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:26:02 localhost systemd[1]: Starting nova_compute container... Nov 27 03:26:02 localhost tripleo-start-podman-container[79898]: Creating additional drop-in dependency for "nova_compute" (ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409) Nov 27 03:26:02 localhost systemd[1]: Reloading. Nov 27 03:26:03 localhost systemd-rc-local-generator[79956]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 03:26:03 localhost systemd-sysv-generator[79961]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 03:26:03 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 03:26:03 localhost systemd[1]: Started nova_compute container. Nov 27 03:26:03 localhost python3[79997]: ansible-file Invoked with path=/var/lib/container-puppet/container-puppet-tasks5.json state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:26:05 localhost python3[80118]: ansible-container_puppet_config Invoked with check_mode=False config_vol_prefix=/var/lib/config-data debug=True net_host=True no_archive=True puppet_config=/var/lib/container-puppet/container-puppet-tasks5.json short_hostname=np0005537446 step=5 update_config_hash_only=False Nov 27 03:26:05 localhost python3[80134]: ansible-file Invoked with path=/var/log/containers/stdouts state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 03:26:05 localhost python3[80150]: ansible-container_config_data Invoked with config_path=/var/lib/tripleo-config/container-puppet-config/step_5 config_pattern=container-puppet-*.json config_overrides={} debug=True Nov 27 03:26:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:26:09 localhost podman[80151]: 2025-11-27 08:26:09.99111979 +0000 UTC m=+0.087938635 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, url=https://www.redhat.com, batch=17.1_20251118.1, config_id=tripleo_step1, build-date=2025-11-18T22:49:46Z, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:26:10 localhost podman[80151]: 2025-11-27 08:26:10.187888832 +0000 UTC m=+0.284707597 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, tcib_managed=true, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, vcs-type=git, distribution-scope=public, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:26:10 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:26:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:26:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:26:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:26:14 localhost systemd[1]: tmp-crun.dVPW8J.mount: Deactivated successfully. Nov 27 03:26:15 localhost podman[80181]: 2025-11-27 08:26:15.001414804 +0000 UTC m=+0.094356049 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, managed_by=tripleo_ansible, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:26:15 localhost podman[80182]: 2025-11-27 08:26:15.007439915 +0000 UTC m=+0.094209074 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, architecture=x86_64, url=https://www.redhat.com) Nov 27 03:26:15 localhost podman[80183]: 2025-11-27 08:26:15.056031702 +0000 UTC m=+0.140678014 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, batch=17.1_20251118.1, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, build-date=2025-11-19T00:11:48Z, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64) Nov 27 03:26:15 localhost podman[80181]: 2025-11-27 08:26:15.106394456 +0000 UTC m=+0.199335731 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, release=1761123044, vcs-type=git, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:26:15 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:26:15 localhost podman[80183]: 2025-11-27 08:26:15.145029425 +0000 UTC m=+0.229675727 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, tcib_managed=true, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, vcs-type=git, container_name=ceilometer_agent_compute, config_id=tripleo_step4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.openshift.expose-services=, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:26:15 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:26:15 localhost podman[80182]: 2025-11-27 08:26:15.161443616 +0000 UTC m=+0.248212705 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, config_id=tripleo_step4, architecture=x86_64, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, io.buildah.version=1.41.4, io.openshift.expose-services=, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:26:15 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:26:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:26:16 localhost podman[80254]: 2025-11-27 08:26:16.978898246 +0000 UTC m=+0.078212584 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, architecture=x86_64, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, vcs-type=git, maintainer=OpenStack TripleO Team) Nov 27 03:26:17 localhost podman[80254]: 2025-11-27 08:26:17.364237727 +0000 UTC m=+0.463552065 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target, release=1761123044, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, tcib_managed=true, build-date=2025-11-19T00:36:58Z, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, version=17.1.12, io.buildah.version=1.41.4) Nov 27 03:26:17 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:26:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:26:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:26:19 localhost systemd[1]: tmp-crun.YcTl3V.mount: Deactivated successfully. Nov 27 03:26:20 localhost podman[80277]: 2025-11-27 08:26:19.996696143 +0000 UTC m=+0.092641752 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, build-date=2025-11-19T00:14:25Z, config_id=tripleo_step4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, container_name=ovn_metadata_agent, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.buildah.version=1.41.4, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 03:26:20 localhost podman[80278]: 2025-11-27 08:26:20.093633489 +0000 UTC m=+0.184468161 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, batch=17.1_20251118.1, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, config_id=tripleo_step4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, container_name=ovn_controller, maintainer=OpenStack TripleO Team, architecture=x86_64, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc.) Nov 27 03:26:20 localhost podman[80277]: 2025-11-27 08:26:20.114700325 +0000 UTC m=+0.210645924 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, managed_by=tripleo_ansible, io.openshift.expose-services=, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, batch=17.1_20251118.1, build-date=2025-11-19T00:14:25Z, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:26:20 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:26:20 localhost podman[80278]: 2025-11-27 08:26:20.141864706 +0000 UTC m=+0.232699348 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, release=1761123044, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_id=tripleo_step4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, version=17.1.12, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z) Nov 27 03:26:20 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:26:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:26:21 localhost systemd[1]: tmp-crun.Q82ECt.mount: Deactivated successfully. Nov 27 03:26:21 localhost podman[80325]: 2025-11-27 08:26:21.985293144 +0000 UTC m=+0.086789885 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., container_name=nova_compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, release=1761123044, url=https://www.redhat.com, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 03:26:22 localhost podman[80325]: 2025-11-27 08:26:22.03792621 +0000 UTC m=+0.139422941 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, architecture=x86_64, config_id=tripleo_step5, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, version=17.1.12, batch=17.1_20251118.1, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=) Nov 27 03:26:22 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:26:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:26:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:26:29 localhost podman[80353]: 2025-11-27 08:26:29.975690719 +0000 UTC m=+0.071479942 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, vendor=Red Hat, Inc., vcs-type=git, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, architecture=x86_64, config_id=tripleo_step3, distribution-scope=public, batch=17.1_20251118.1, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z) Nov 27 03:26:29 localhost podman[80353]: 2025-11-27 08:26:29.986257504 +0000 UTC m=+0.082046727 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, tcib_managed=true, version=17.1.12, managed_by=tripleo_ansible, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, container_name=iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public) Nov 27 03:26:29 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:26:30 localhost systemd[1]: tmp-crun.V02fee.mount: Deactivated successfully. Nov 27 03:26:30 localhost podman[80352]: 2025-11-27 08:26:30.036448673 +0000 UTC m=+0.134991490 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, batch=17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vcs-type=git, com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, container_name=collectd) Nov 27 03:26:30 localhost podman[80352]: 2025-11-27 08:26:30.04786761 +0000 UTC m=+0.146410447 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, vendor=Red Hat, Inc., distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, version=17.1.12, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, architecture=x86_64, tcib_managed=true, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, config_id=tripleo_step3) Nov 27 03:26:30 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:26:36 localhost sshd[80392]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:26:36 localhost systemd-logind[761]: New session 33 of user zuul. Nov 27 03:26:36 localhost systemd[1]: Started Session 33 of User zuul. Nov 27 03:26:37 localhost python3[80501]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 03:26:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:26:40 localhost podman[80688]: 2025-11-27 08:26:40.985465783 +0000 UTC m=+0.081596485 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-type=git, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, distribution-scope=public, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1) Nov 27 03:26:41 localhost podman[80688]: 2025-11-27 08:26:41.168223388 +0000 UTC m=+0.264353980 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_id=tripleo_step1, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-type=git, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, tcib_managed=true, container_name=metrics_qdr, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12) Nov 27 03:26:41 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:26:44 localhost python3[80792]: ansible-ansible.legacy.dnf Invoked with name=['iptables'] allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None state=None Nov 27 03:26:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:26:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:26:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:26:45 localhost systemd[1]: tmp-crun.MGlhA3.mount: Deactivated successfully. Nov 27 03:26:45 localhost podman[80796]: 2025-11-27 08:26:45.992876168 +0000 UTC m=+0.084751790 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, distribution-scope=public, managed_by=tripleo_ansible, url=https://www.redhat.com, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:26:46 localhost podman[80796]: 2025-11-27 08:26:46.031147007 +0000 UTC m=+0.123022609 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, container_name=logrotate_crond, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, vendor=Red Hat, Inc., tcib_managed=true, config_id=tripleo_step4, name=rhosp17/openstack-cron) Nov 27 03:26:46 localhost systemd[1]: tmp-crun.4lV3kS.mount: Deactivated successfully. Nov 27 03:26:46 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:26:46 localhost podman[80795]: 2025-11-27 08:26:46.059313255 +0000 UTC m=+0.152237896 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, maintainer=OpenStack TripleO Team, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, vendor=Red Hat, Inc., io.buildah.version=1.41.4, url=https://www.redhat.com, architecture=x86_64, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:26:46 localhost podman[80795]: 2025-11-27 08:26:46.090947625 +0000 UTC m=+0.183872286 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 03:26:46 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:26:46 localhost podman[80797]: 2025-11-27 08:26:46.152930822 +0000 UTC m=+0.240173960 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, io.openshift.expose-services=, distribution-scope=public, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:26:46 localhost podman[80797]: 2025-11-27 08:26:46.205915296 +0000 UTC m=+0.293158464 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, managed_by=tripleo_ansible, distribution-scope=public, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:26:46 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:26:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:26:48 localhost systemd[1]: tmp-crun.DB7j2B.mount: Deactivated successfully. Nov 27 03:26:48 localhost podman[80939]: 2025-11-27 08:26:48.043867277 +0000 UTC m=+0.133664155 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_id=tripleo_step4, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, distribution-scope=public, url=https://www.redhat.com, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:26:48 localhost python3[80972]: ansible-ansible.builtin.iptables Invoked with action=insert chain=INPUT comment=allow ssh access for zuul executor in_interface=eth0 jump=ACCEPT protocol=tcp source=38.102.83.114 table=filter state=present ip_version=ipv4 match=[] destination_ports=[] ctstate=[] syn=ignore flush=False chain_management=False numeric=False rule_num=None wait=None to_source=None destination=None to_destination=None tcp_flags=None gateway=None log_prefix=None log_level=None goto=None out_interface=None fragment=None set_counters=None source_port=None destination_port=None to_ports=None set_dscp_mark=None set_dscp_mark_class=None src_range=None dst_range=None match_set=None match_set_flags=None limit=None limit_burst=None uid_owner=None gid_owner=None reject_with=None icmp_type=None policy=None Nov 27 03:26:48 localhost kernel: Warning: Deprecated Driver is detected: nft_compat will not be maintained in a future major release and may be disabled Nov 27 03:26:48 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 81.1 (270 of 333 items), suggesting rotation. Nov 27 03:26:48 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 03:26:48 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 03:26:48 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 03:26:48 localhost podman[80939]: 2025-11-27 08:26:48.413235639 +0000 UTC m=+0.503032507 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, version=17.1.12, config_id=tripleo_step4, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, io.openshift.expose-services=, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:26:48 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:26:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:26:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:26:50 localhost systemd[1]: tmp-crun.0sZ9pc.mount: Deactivated successfully. Nov 27 03:26:51 localhost podman[81048]: 2025-11-27 08:26:51.000091037 +0000 UTC m=+0.094922073 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, container_name=ovn_controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, batch=17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, com.redhat.component=openstack-ovn-controller-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:26:51 localhost podman[81048]: 2025-11-27 08:26:51.02586226 +0000 UTC m=+0.120693376 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, vendor=Red Hat, Inc., distribution-scope=public, build-date=2025-11-18T23:34:05Z, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, tcib_managed=true, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4) Nov 27 03:26:51 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:26:51 localhost podman[81047]: 2025-11-27 08:26:51.045387285 +0000 UTC m=+0.140521819 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, tcib_managed=true, architecture=x86_64, url=https://www.redhat.com, container_name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:26:51 localhost podman[81047]: 2025-11-27 08:26:51.120084704 +0000 UTC m=+0.215219198 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, config_id=tripleo_step4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., tcib_managed=true, version=17.1.12, distribution-scope=public, batch=17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64) Nov 27 03:26:51 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:26:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:26:53 localhost podman[81155]: 2025-11-27 08:26:53.032200809 +0000 UTC m=+0.119790622 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, architecture=x86_64, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, batch=17.1_20251118.1, distribution-scope=public, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, url=https://www.redhat.com) Nov 27 03:26:53 localhost podman[81155]: 2025-11-27 08:26:53.089865499 +0000 UTC m=+0.177455282 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_compute, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, config_id=tripleo_step5, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, vendor=Red Hat, Inc., architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, release=1761123044, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4) Nov 27 03:26:53 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:27:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:27:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:27:00 localhost podman[81197]: 2025-11-27 08:27:00.987187881 +0000 UTC m=+0.082911600 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, managed_by=tripleo_ansible, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, distribution-scope=public, build-date=2025-11-18T23:44:13Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git) Nov 27 03:27:01 localhost podman[81197]: 2025-11-27 08:27:01.001829015 +0000 UTC m=+0.097552714 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, config_id=tripleo_step3, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z, container_name=iscsid, tcib_managed=true, batch=17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:27:01 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:27:01 localhost systemd[1]: tmp-crun.VoSlOl.mount: Deactivated successfully. Nov 27 03:27:01 localhost podman[81196]: 2025-11-27 08:27:01.057980085 +0000 UTC m=+0.155896623 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, vcs-type=git, architecture=x86_64, build-date=2025-11-18T22:51:28Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, batch=17.1_20251118.1) Nov 27 03:27:01 localhost podman[81196]: 2025-11-27 08:27:01.095711519 +0000 UTC m=+0.193628067 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, vcs-type=git, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, architecture=x86_64, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, version=17.1.12, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:27:01 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:27:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:27:11 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:27:11 localhost recover_tripleo_nova_virtqemud[81242]: 63639 Nov 27 03:27:11 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:27:11 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:27:11 localhost podman[81235]: 2025-11-27 08:27:11.991555179 +0000 UTC m=+0.088095711 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, batch=17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd) Nov 27 03:27:12 localhost podman[81235]: 2025-11-27 08:27:12.216855097 +0000 UTC m=+0.313395659 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, tcib_managed=true, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., managed_by=tripleo_ansible, url=https://www.redhat.com, architecture=x86_64, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:27:12 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:27:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:27:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:27:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:27:17 localhost systemd[1]: tmp-crun.pauWM7.mount: Deactivated successfully. Nov 27 03:27:17 localhost podman[81266]: 2025-11-27 08:27:17.006438776 +0000 UTC m=+0.095595241 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, url=https://www.redhat.com, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, tcib_managed=true, distribution-scope=public, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, architecture=x86_64, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc.) Nov 27 03:27:17 localhost systemd[1]: tmp-crun.ndZgoj.mount: Deactivated successfully. Nov 27 03:27:17 localhost podman[81267]: 2025-11-27 08:27:17.05571003 +0000 UTC m=+0.145678278 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, distribution-scope=public, container_name=logrotate_crond, config_id=tripleo_step4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-cron, tcib_managed=true) Nov 27 03:27:17 localhost podman[81266]: 2025-11-27 08:27:17.067858448 +0000 UTC m=+0.157014883 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, architecture=x86_64, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1) Nov 27 03:27:17 localhost podman[81268]: 2025-11-27 08:27:17.114142862 +0000 UTC m=+0.200705248 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, version=17.1.12, batch=17.1_20251118.1, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, release=1761123044, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:27:17 localhost podman[81267]: 2025-11-27 08:27:17.118311074 +0000 UTC m=+0.208279292 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, tcib_managed=true, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, version=17.1.12, release=1761123044, vcs-type=git, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:27:17 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:27:17 localhost podman[81268]: 2025-11-27 08:27:17.147960091 +0000 UTC m=+0.234522477 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_id=tripleo_step4, container_name=ceilometer_agent_compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., version=17.1.12, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:27:17 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:27:17 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:27:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:27:19 localhost podman[81341]: 2025-11-27 08:27:19.001380608 +0000 UTC m=+0.088474250 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, architecture=x86_64, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vcs-type=git, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, batch=17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, tcib_managed=true, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:27:19 localhost podman[81341]: 2025-11-27 08:27:19.390933693 +0000 UTC m=+0.478027315 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, version=17.1.12, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:27:19 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:27:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:27:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:27:21 localhost podman[81366]: 2025-11-27 08:27:21.988399486 +0000 UTC m=+0.079160899 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, architecture=x86_64, tcib_managed=true, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible) Nov 27 03:27:22 localhost podman[81366]: 2025-11-27 08:27:22.009137054 +0000 UTC m=+0.099898427 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-type=git, managed_by=tripleo_ansible, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc., distribution-scope=public, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:27:22 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:27:22 localhost systemd[1]: tmp-crun.4i0jXs.mount: Deactivated successfully. Nov 27 03:27:22 localhost podman[81365]: 2025-11-27 08:27:22.108790614 +0000 UTC m=+0.200548804 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack TripleO Team, version=17.1.12, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, vendor=Red Hat, Inc.) Nov 27 03:27:22 localhost podman[81365]: 2025-11-27 08:27:22.155028947 +0000 UTC m=+0.246787127 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 03:27:22 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:27:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:27:23 localhost systemd[1]: tmp-crun.KpcCKn.mount: Deactivated successfully. Nov 27 03:27:23 localhost podman[81413]: 2025-11-27 08:27:23.9941474 +0000 UTC m=+0.087565685 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, io.openshift.expose-services=, vcs-type=git, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, maintainer=OpenStack TripleO Team, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:27:24 localhost podman[81413]: 2025-11-27 08:27:24.029191573 +0000 UTC m=+0.122609848 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, url=https://www.redhat.com, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:27:24 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:27:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:27:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:27:31 localhost systemd[1]: tmp-crun.GPOa3j.mount: Deactivated successfully. Nov 27 03:27:31 localhost podman[81439]: 2025-11-27 08:27:31.997592096 +0000 UTC m=+0.087457122 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, tcib_managed=true, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, container_name=iscsid, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:27:32 localhost podman[81439]: 2025-11-27 08:27:32.039840792 +0000 UTC m=+0.129705738 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, build-date=2025-11-18T23:44:13Z, vcs-type=git, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, container_name=iscsid, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container) Nov 27 03:27:32 localhost podman[81438]: 2025-11-27 08:27:32.047160369 +0000 UTC m=+0.138792743 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, config_id=tripleo_step3, container_name=collectd, name=rhosp17/openstack-collectd, distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, build-date=2025-11-18T22:51:28Z, release=1761123044, version=17.1.12, io.openshift.expose-services=) Nov 27 03:27:32 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:27:32 localhost podman[81438]: 2025-11-27 08:27:32.058782222 +0000 UTC m=+0.150414596 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, build-date=2025-11-18T22:51:28Z, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, release=1761123044, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, version=17.1.12) Nov 27 03:27:32 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:27:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:27:42 localhost podman[81479]: 2025-11-27 08:27:42.994253267 +0000 UTC m=+0.089595981 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, tcib_managed=true, version=17.1.12, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044) Nov 27 03:27:43 localhost podman[81479]: 2025-11-27 08:27:43.209856304 +0000 UTC m=+0.305198978 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, io.openshift.expose-services=, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, vendor=Red Hat, Inc., release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, vcs-type=git, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:27:43 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:27:47 localhost systemd[1]: session-33.scope: Deactivated successfully. Nov 27 03:27:47 localhost systemd[1]: session-33.scope: Consumed 5.610s CPU time. Nov 27 03:27:47 localhost systemd-logind[761]: Session 33 logged out. Waiting for processes to exit. Nov 27 03:27:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:27:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:27:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:27:47 localhost systemd-logind[761]: Removed session 33. Nov 27 03:27:47 localhost systemd[1]: tmp-crun.ir637e.mount: Deactivated successfully. Nov 27 03:27:47 localhost podman[81508]: 2025-11-27 08:27:47.590885857 +0000 UTC m=+0.143898820 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, release=1761123044, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:12:45Z) Nov 27 03:27:47 localhost podman[81509]: 2025-11-27 08:27:47.641444376 +0000 UTC m=+0.189488245 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, version=17.1.12, release=1761123044, tcib_managed=true, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, batch=17.1_20251118.1, architecture=x86_64, config_id=tripleo_step4, io.openshift.expose-services=, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z) Nov 27 03:27:47 localhost podman[81510]: 2025-11-27 08:27:47.56011566 +0000 UTC m=+0.107697567 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, distribution-scope=public, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, build-date=2025-11-19T00:11:48Z, architecture=x86_64, version=17.1.12, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:27:47 localhost podman[81509]: 2025-11-27 08:27:47.676458718 +0000 UTC m=+0.224502607 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-type=git, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, batch=17.1_20251118.1, container_name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, maintainer=OpenStack TripleO Team) Nov 27 03:27:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:27:47 localhost podman[81510]: 2025-11-27 08:27:47.694950025 +0000 UTC m=+0.242531902 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, tcib_managed=true, version=17.1.12, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, batch=17.1_20251118.1, release=1761123044, io.openshift.expose-services=) Nov 27 03:27:47 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:27:47 localhost podman[81508]: 2025-11-27 08:27:47.751414614 +0000 UTC m=+0.304427587 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, tcib_managed=true, version=17.1.12, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-type=git, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:27:47 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:27:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:27:49 localhost podman[81580]: 2025-11-27 08:27:49.982357762 +0000 UTC m=+0.080505676 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, distribution-scope=public, vendor=Red Hat, Inc., managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, tcib_managed=true, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.buildah.version=1.41.4) Nov 27 03:27:50 localhost podman[81580]: 2025-11-27 08:27:50.358916407 +0000 UTC m=+0.457064311 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., config_id=tripleo_step4, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:27:50 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:27:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:27:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 2400.1 total, 600.0 interval#012Cumulative writes: 4601 writes, 21K keys, 4601 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4601 writes, 459 syncs, 10.02 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:27:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:27:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:27:52 localhost podman[81649]: 2025-11-27 08:27:52.989675446 +0000 UTC m=+0.082381146 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., config_id=tripleo_step4, managed_by=tripleo_ansible, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, distribution-scope=public, release=1761123044, io.buildah.version=1.41.4, vcs-type=git, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12) Nov 27 03:27:53 localhost podman[81649]: 2025-11-27 08:27:53.058267301 +0000 UTC m=+0.150972991 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.expose-services=, version=17.1.12, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, distribution-scope=public, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:27:53 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:27:53 localhost podman[81650]: 2025-11-27 08:27:53.062732032 +0000 UTC m=+0.149801930 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, name=rhosp17/openstack-ovn-controller, version=17.1.12, container_name=ovn_controller, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., vcs-type=git, io.buildah.version=1.41.4, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, architecture=x86_64, batch=17.1_20251118.1, config_id=tripleo_step4, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:27:53 localhost podman[81650]: 2025-11-27 08:27:53.142465035 +0000 UTC m=+0.229534943 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, url=https://www.redhat.com, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, build-date=2025-11-18T23:34:05Z, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git) Nov 27 03:27:53 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:27:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:27:55 localhost systemd[1]: tmp-crun.C2VeUk.mount: Deactivated successfully. Nov 27 03:27:55 localhost podman[81760]: 2025-11-27 08:27:55.035085996 +0000 UTC m=+0.113208046 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, distribution-scope=public, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, container_name=nova_compute, architecture=x86_64, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., version=17.1.12, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:27:55 localhost podman[81760]: 2025-11-27 08:27:55.070839476 +0000 UTC m=+0.148961506 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-type=git, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, version=17.1.12, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:27:55 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:27:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:27:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 2400.2 total, 600.0 interval#012Cumulative writes: 4903 writes, 21K keys, 4903 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 4903 writes, 543 syncs, 9.03 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:28:01 localhost sshd[81801]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:28:01 localhost systemd-logind[761]: New session 34 of user zuul. Nov 27 03:28:01 localhost systemd[1]: Started Session 34 of User zuul. Nov 27 03:28:02 localhost python3[81820]: ansible-ansible.legacy.dnf Invoked with name=['systemd-container'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 03:28:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:28:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:28:03 localhost podman[81822]: 2025-11-27 08:28:03.007100436 +0000 UTC m=+0.097030530 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, config_id=tripleo_step3, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, container_name=collectd, release=1761123044, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, architecture=x86_64, vendor=Red Hat, Inc., managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:28:03 localhost podman[81822]: 2025-11-27 08:28:03.05001777 +0000 UTC m=+0.139947834 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., url=https://www.redhat.com, io.openshift.expose-services=, vcs-type=git, batch=17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, container_name=collectd, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container) Nov 27 03:28:03 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:28:03 localhost podman[81823]: 2025-11-27 08:28:03.065199498 +0000 UTC m=+0.153644722 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, url=https://www.redhat.com, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, architecture=x86_64, io.buildah.version=1.41.4, release=1761123044, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, container_name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:28:03 localhost podman[81823]: 2025-11-27 08:28:03.069819163 +0000 UTC m=+0.158264417 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, vcs-type=git, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., container_name=iscsid, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, distribution-scope=public) Nov 27 03:28:03 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:28:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:28:13 localhost podman[81860]: 2025-11-27 08:28:13.992857693 +0000 UTC m=+0.086532418 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, url=https://www.redhat.com, architecture=x86_64, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, container_name=metrics_qdr, vendor=Red Hat, Inc., config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:28:14 localhost podman[81860]: 2025-11-27 08:28:14.181473504 +0000 UTC m=+0.275148229 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, version=17.1.12, vcs-type=git, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, release=1761123044, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:28:14 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:28:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:28:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:28:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:28:18 localhost systemd[1]: tmp-crun.TOLzTc.mount: Deactivated successfully. Nov 27 03:28:18 localhost podman[81889]: 2025-11-27 08:28:18.020218846 +0000 UTC m=+0.103937427 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, config_id=tripleo_step4, io.buildah.version=1.41.4, release=1761123044, managed_by=tripleo_ansible, tcib_managed=true, version=17.1.12, url=https://www.redhat.com) Nov 27 03:28:18 localhost podman[81891]: 2025-11-27 08:28:18.076788847 +0000 UTC m=+0.151955418 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, vendor=Red Hat, Inc., container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, release=1761123044, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_id=tripleo_step4, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.openshift.expose-services=, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-compute-container, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:28:18 localhost podman[81891]: 2025-11-27 08:28:18.108926921 +0000 UTC m=+0.184093512 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, release=1761123044, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., batch=17.1_20251118.1, container_name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute) Nov 27 03:28:18 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:28:18 localhost podman[81890]: 2025-11-27 08:28:18.168907284 +0000 UTC m=+0.248862473 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, tcib_managed=true, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, vcs-type=git, version=17.1.12, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, container_name=logrotate_crond) Nov 27 03:28:18 localhost podman[81890]: 2025-11-27 08:28:18.207967274 +0000 UTC m=+0.287922453 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, version=17.1.12, tcib_managed=true, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_id=tripleo_step4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044) Nov 27 03:28:18 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:28:18 localhost podman[81889]: 2025-11-27 08:28:18.261436272 +0000 UTC m=+0.345154813 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-type=git, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vendor=Red Hat, Inc., io.buildah.version=1.41.4, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:28:18 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:28:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:28:21 localhost podman[81963]: 2025-11-27 08:28:21.007951394 +0000 UTC m=+0.105945620 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:28:21 localhost podman[81963]: 2025-11-27 08:28:21.4023863 +0000 UTC m=+0.500380566 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-type=git, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, release=1761123044) Nov 27 03:28:21 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:28:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:28:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:28:24 localhost systemd[1]: tmp-crun.IlSKyd.mount: Deactivated successfully. Nov 27 03:28:24 localhost podman[81985]: 2025-11-27 08:28:24.007610093 +0000 UTC m=+0.093979079 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:28:24 localhost systemd[1]: tmp-crun.nprnss.mount: Deactivated successfully. Nov 27 03:28:24 localhost podman[81984]: 2025-11-27 08:28:24.063684131 +0000 UTC m=+0.153715795 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, tcib_managed=true, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, release=1761123044, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:28:24 localhost podman[81985]: 2025-11-27 08:28:24.08039463 +0000 UTC m=+0.166763616 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, release=1761123044, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, vendor=Red Hat, Inc.) Nov 27 03:28:24 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:28:24 localhost podman[81984]: 2025-11-27 08:28:24.136732735 +0000 UTC m=+0.226764449 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, version=17.1.12, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:28:24 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:28:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:28:25 localhost systemd[1]: tmp-crun.U9BMmF.mount: Deactivated successfully. Nov 27 03:28:25 localhost podman[82030]: 2025-11-27 08:28:25.987739628 +0000 UTC m=+0.081375460 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, name=rhosp17/openstack-nova-compute, version=17.1.12, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vendor=Red Hat, Inc., container_name=nova_compute, architecture=x86_64, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team) Nov 27 03:28:26 localhost podman[82030]: 2025-11-27 08:28:26.012978456 +0000 UTC m=+0.106614268 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, vcs-type=git, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, config_id=tripleo_step5, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, version=17.1.12, tcib_managed=true) Nov 27 03:28:26 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:28:31 localhost python3[82071]: ansible-ansible.legacy.dnf Invoked with name=['sos'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 27 03:28:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:28:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:28:34 localhost podman[82075]: 2025-11-27 08:28:34.023099133 +0000 UTC m=+0.112219493 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, batch=17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:28:34 localhost podman[82075]: 2025-11-27 08:28:34.063844904 +0000 UTC m=+0.152965224 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, vendor=Red Hat, Inc., container_name=collectd, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:28:34 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:28:34 localhost podman[82076]: 2025-11-27 08:28:33.985743884 +0000 UTC m=+0.075156772 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, batch=17.1_20251118.1, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, release=1761123044, architecture=x86_64, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, vendor=Red Hat, Inc., url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:28:34 localhost podman[82076]: 2025-11-27 08:28:34.122950751 +0000 UTC m=+0.212363839 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, build-date=2025-11-18T23:44:13Z, tcib_managed=true, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, vcs-type=git, version=17.1.12, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, architecture=x86_64, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:28:34 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:28:34 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 03:28:34 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 03:28:34 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 03:28:35 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 03:28:35 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 03:28:35 localhost systemd[1]: run-r2788ae076e014e1abc353a2ffd7a7972.service: Deactivated successfully. Nov 27 03:28:35 localhost systemd[1]: run-rafbe2d4ba9a942938f5693ae397af7bc.service: Deactivated successfully. Nov 27 03:28:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:28:44 localhost systemd[1]: tmp-crun.nef9uM.mount: Deactivated successfully. Nov 27 03:28:44 localhost podman[82261]: 2025-11-27 08:28:44.973864916 +0000 UTC m=+0.070546867 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, vcs-type=git, distribution-scope=public, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, maintainer=OpenStack TripleO Team) Nov 27 03:28:45 localhost podman[82261]: 2025-11-27 08:28:45.18204462 +0000 UTC m=+0.278726571 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, release=1761123044, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:28:45 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:28:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:28:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:28:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:28:49 localhost systemd[1]: tmp-crun.aNGWd0.mount: Deactivated successfully. Nov 27 03:28:49 localhost podman[82289]: 2025-11-27 08:28:49.014194353 +0000 UTC m=+0.094845814 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, distribution-scope=public, vcs-type=git, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vendor=Red Hat, Inc., managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:28:49 localhost podman[82290]: 2025-11-27 08:28:49.064468941 +0000 UTC m=+0.142041478 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, config_id=tripleo_step4, io.openshift.expose-services=, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, tcib_managed=true, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=logrotate_crond) Nov 27 03:28:49 localhost podman[82290]: 2025-11-27 08:28:49.073154066 +0000 UTC m=+0.150726623 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, version=17.1.12, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, distribution-scope=public, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z) Nov 27 03:28:49 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:28:49 localhost podman[82289]: 2025-11-27 08:28:49.123791254 +0000 UTC m=+0.204442725 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, release=1761123044, architecture=x86_64, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, tcib_managed=true, io.openshift.expose-services=, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:28:49 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:28:49 localhost podman[82291]: 2025-11-27 08:28:49.21391747 +0000 UTC m=+0.292338511 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, container_name=ceilometer_agent_compute, version=17.1.12, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, vcs-type=git, build-date=2025-11-19T00:11:48Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., url=https://www.redhat.com, config_id=tripleo_step4, tcib_managed=true, distribution-scope=public, io.buildah.version=1.41.4, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:28:49 localhost podman[82291]: 2025-11-27 08:28:49.266259513 +0000 UTC m=+0.344680574 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, vendor=Red Hat, Inc., vcs-type=git, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64) Nov 27 03:28:49 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:28:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:28:51 localhost podman[82362]: 2025-11-27 08:28:51.979913375 +0000 UTC m=+0.079225351 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.openshift.expose-services=, distribution-scope=public, io.buildah.version=1.41.4, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-type=git, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:28:52 localhost podman[82362]: 2025-11-27 08:28:52.383058828 +0000 UTC m=+0.482370854 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, distribution-scope=public, tcib_managed=true, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:28:52 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:28:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:28:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:28:54 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:28:54 localhost recover_tripleo_nova_virtqemud[82443]: 63639 Nov 27 03:28:54 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:28:54 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:28:55 localhost systemd[1]: tmp-crun.B5DGvs.mount: Deactivated successfully. Nov 27 03:28:55 localhost systemd[1]: tmp-crun.O7p9gZ.mount: Deactivated successfully. Nov 27 03:28:55 localhost podman[82432]: 2025-11-27 08:28:55.057088057 +0000 UTC m=+0.148377140 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-19T00:14:25Z, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step4, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:28:55 localhost podman[82433]: 2025-11-27 08:28:55.016970504 +0000 UTC m=+0.105755319 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, architecture=x86_64, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, release=1761123044, build-date=2025-11-18T23:34:05Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4) Nov 27 03:28:55 localhost podman[82433]: 2025-11-27 08:28:55.097399197 +0000 UTC m=+0.186183972 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z) Nov 27 03:28:55 localhost podman[82432]: 2025-11-27 08:28:55.105987709 +0000 UTC m=+0.197276742 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, architecture=x86_64, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team) Nov 27 03:28:55 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:28:55 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:28:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:28:56 localhost systemd[1]: tmp-crun.5ITI5m.mount: Deactivated successfully. Nov 27 03:28:57 localhost podman[82482]: 2025-11-27 08:28:57.006880829 +0000 UTC m=+0.098838851 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, architecture=x86_64, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, container_name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, config_id=tripleo_step5, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vendor=Red Hat, Inc., io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1) Nov 27 03:28:57 localhost podman[82482]: 2025-11-27 08:28:57.061154126 +0000 UTC m=+0.153112168 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step5, distribution-scope=public, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, architecture=x86_64, container_name=nova_compute) Nov 27 03:28:57 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:29:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:29:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:29:05 localhost podman[82635]: 2025-11-27 08:29:04.998704951 +0000 UTC m=+0.091806250 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, distribution-scope=public, container_name=collectd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-type=git, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:29:05 localhost podman[82635]: 2025-11-27 08:29:05.011843797 +0000 UTC m=+0.104945116 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, container_name=collectd, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, release=1761123044, io.openshift.expose-services=) Nov 27 03:29:05 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:29:05 localhost podman[82636]: 2025-11-27 08:29:05.098778116 +0000 UTC m=+0.189667086 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, tcib_managed=true, architecture=x86_64, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.component=openstack-iscsid-container, vcs-type=git, distribution-scope=public, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:29:05 localhost podman[82636]: 2025-11-27 08:29:05.111972062 +0000 UTC m=+0.202860992 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.expose-services=, config_id=tripleo_step3, release=1761123044, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, vcs-type=git, com.redhat.component=openstack-iscsid-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:29:05 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:29:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:29:16 localhost podman[82673]: 2025-11-27 08:29:15.999449892 +0000 UTC m=+0.089863339 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step1, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, architecture=x86_64, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:29:16 localhost podman[82673]: 2025-11-27 08:29:16.225210602 +0000 UTC m=+0.315623969 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, version=17.1.12, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, build-date=2025-11-18T22:49:46Z) Nov 27 03:29:16 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:29:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:29:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:29:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:29:19 localhost podman[82702]: 2025-11-27 08:29:19.986599921 +0000 UTC m=+0.080596718 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., release=1761123044, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:29:20 localhost systemd[1]: tmp-crun.FGnHWb.mount: Deactivated successfully. Nov 27 03:29:20 localhost podman[82703]: 2025-11-27 08:29:20.044081615 +0000 UTC m=+0.134827534 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.buildah.version=1.41.4, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, distribution-scope=public, architecture=x86_64, vcs-type=git, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z) Nov 27 03:29:20 localhost podman[82703]: 2025-11-27 08:29:20.082058191 +0000 UTC m=+0.172804140 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, maintainer=OpenStack TripleO Team, architecture=x86_64, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, tcib_managed=true, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, config_id=tripleo_step4, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:29:20 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:29:20 localhost podman[82704]: 2025-11-27 08:29:20.098753621 +0000 UTC m=+0.188577115 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, build-date=2025-11-19T00:11:48Z, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, managed_by=tripleo_ansible, release=1761123044, version=17.1.12, container_name=ceilometer_agent_compute, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:29:20 localhost podman[82702]: 2025-11-27 08:29:20.116200094 +0000 UTC m=+0.210196891 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, release=1761123044, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_ipmi, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, distribution-scope=public, config_id=tripleo_step4, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:29:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:29:20 localhost podman[82704]: 2025-11-27 08:29:20.161812346 +0000 UTC m=+0.251635800 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, batch=17.1_20251118.1, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, architecture=x86_64, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, vcs-type=git, tcib_managed=true, url=https://www.redhat.com, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, version=17.1.12, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044) Nov 27 03:29:20 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:29:22 localhost python3[82790]: ansible-ansible.legacy.command Invoked with _raw_params=subscription-manager repos --disable rhel-9-for-x86_64-baseos-eus-rpms --disable rhel-9-for-x86_64-appstream-eus-rpms --disable rhel-9-for-x86_64-highavailability-eus-rpms --disable openstack-17.1-for-rhel-9-x86_64-rpms --disable fast-datapath-for-rhel-9-x86_64-rpms _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 03:29:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:29:22 localhost systemd[1]: tmp-crun.XRTZqE.mount: Deactivated successfully. Nov 27 03:29:22 localhost podman[82793]: 2025-11-27 08:29:22.983674941 +0000 UTC m=+0.079376006 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, version=17.1.12, distribution-scope=public, container_name=nova_migration_target, architecture=x86_64, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, config_id=tripleo_step4, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:29:23 localhost podman[82793]: 2025-11-27 08:29:23.378700344 +0000 UTC m=+0.474401419 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-type=git, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute, release=1761123044, batch=17.1_20251118.1) Nov 27 03:29:23 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:29:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:29:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:29:25 localhost podman[82938]: 2025-11-27 08:29:25.992669791 +0000 UTC m=+0.086548238 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ovn-controller, distribution-scope=public, build-date=2025-11-18T23:34:05Z, release=1761123044, version=17.1.12, batch=17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, container_name=ovn_controller, io.buildah.version=1.41.4) Nov 27 03:29:26 localhost podman[82938]: 2025-11-27 08:29:26.020706389 +0000 UTC m=+0.114584816 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, distribution-scope=public, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, tcib_managed=true, build-date=2025-11-18T23:34:05Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git) Nov 27 03:29:26 localhost systemd[1]: tmp-crun.k8XRhA.mount: Deactivated successfully. Nov 27 03:29:26 localhost podman[82937]: 2025-11-27 08:29:26.041926943 +0000 UTC m=+0.139832850 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., architecture=x86_64, tcib_managed=true, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.openshift.expose-services=, version=17.1.12, config_id=tripleo_step4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}) Nov 27 03:29:26 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:29:26 localhost podman[82937]: 2025-11-27 08:29:26.116851697 +0000 UTC m=+0.214757594 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_id=tripleo_step4, vcs-type=git, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, tcib_managed=true, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:29:26 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:29:26 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 03:29:26 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 03:29:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:29:27 localhost podman[82992]: 2025-11-27 08:29:27.991059037 +0000 UTC m=+0.087575897 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, vcs-type=git, io.openshift.expose-services=, batch=17.1_20251118.1, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, version=17.1.12, tcib_managed=true, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4) Nov 27 03:29:28 localhost podman[82992]: 2025-11-27 08:29:28.049932418 +0000 UTC m=+0.146449328 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., architecture=x86_64, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, tcib_managed=true, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, config_id=tripleo_step5, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:29:28 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:29:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:29:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:29:35 localhost podman[83078]: 2025-11-27 08:29:35.989691785 +0000 UTC m=+0.086858368 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, version=17.1.12, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, distribution-scope=public, config_id=tripleo_step3, container_name=collectd, vcs-type=git, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:29:36 localhost podman[83079]: 2025-11-27 08:29:36.036465099 +0000 UTC m=+0.131618797 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, vcs-type=git, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, release=1761123044, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:29:36 localhost podman[83078]: 2025-11-27 08:29:36.055175304 +0000 UTC m=+0.152341847 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, batch=17.1_20251118.1, vendor=Red Hat, Inc., container_name=collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, name=rhosp17/openstack-collectd, io.openshift.expose-services=, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3) Nov 27 03:29:36 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:29:36 localhost podman[83079]: 2025-11-27 08:29:36.070062727 +0000 UTC m=+0.165216445 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, version=17.1.12, architecture=x86_64, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, container_name=iscsid, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, vcs-type=git, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1) Nov 27 03:29:36 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:29:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:29:46 localhost systemd[1]: tmp-crun.libo20.mount: Deactivated successfully. Nov 27 03:29:46 localhost podman[83116]: 2025-11-27 08:29:46.998799873 +0000 UTC m=+0.094483554 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, tcib_managed=true, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-type=git, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=) Nov 27 03:29:47 localhost podman[83116]: 2025-11-27 08:29:47.233647468 +0000 UTC m=+0.329331179 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, container_name=metrics_qdr, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, version=17.1.12, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z) Nov 27 03:29:47 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:29:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:29:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:29:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:29:50 localhost systemd[1]: tmp-crun.bdwi18.mount: Deactivated successfully. Nov 27 03:29:51 localhost podman[83144]: 2025-11-27 08:29:51.003716144 +0000 UTC m=+0.099146471 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, com.redhat.component=openstack-ceilometer-ipmi-container, tcib_managed=true, build-date=2025-11-19T00:12:45Z, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., release=1761123044, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:29:51 localhost podman[83145]: 2025-11-27 08:29:51.044065233 +0000 UTC m=+0.135775860 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, tcib_managed=true, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, build-date=2025-11-18T22:49:32Z, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, release=1761123044, io.openshift.expose-services=, batch=17.1_20251118.1, container_name=logrotate_crond, vcs-type=git, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:29:51 localhost podman[83145]: 2025-11-27 08:29:51.051645689 +0000 UTC m=+0.143356306 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, architecture=x86_64, container_name=logrotate_crond, release=1761123044, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:29:51 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:29:51 localhost podman[83146]: 2025-11-27 08:29:51.10537074 +0000 UTC m=+0.193570351 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, config_id=tripleo_step4, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true, distribution-scope=public, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-compute) Nov 27 03:29:51 localhost podman[83144]: 2025-11-27 08:29:51.111264889 +0000 UTC m=+0.206695226 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, tcib_managed=true, release=1761123044, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:29:51 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:29:51 localhost podman[83146]: 2025-11-27 08:29:51.138890856 +0000 UTC m=+0.227090437 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, release=1761123044, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, version=17.1.12, config_id=tripleo_step4, tcib_managed=true, distribution-scope=public, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z) Nov 27 03:29:51 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:29:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:29:53 localhost podman[83262]: 2025-11-27 08:29:53.989158087 +0000 UTC m=+0.083461276 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, config_id=tripleo_step4, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-type=git, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, tcib_managed=true, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:29:54 localhost podman[83262]: 2025-11-27 08:29:54.361674392 +0000 UTC m=+0.455977651 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=) Nov 27 03:29:54 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:29:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:29:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:29:56 localhost podman[83285]: 2025-11-27 08:29:56.984180681 +0000 UTC m=+0.082026127 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, build-date=2025-11-19T00:14:25Z, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, release=1761123044, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git) Nov 27 03:29:57 localhost podman[83286]: 2025-11-27 08:29:57.03781893 +0000 UTC m=+0.133780065 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, container_name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, tcib_managed=true, name=rhosp17/openstack-ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, vcs-type=git, io.buildah.version=1.41.4, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044) Nov 27 03:29:57 localhost podman[83285]: 2025-11-27 08:29:57.057023389 +0000 UTC m=+0.154868875 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, config_id=tripleo_step4, io.buildah.version=1.41.4, url=https://www.redhat.com, architecture=x86_64, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:29:57 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:29:57 localhost podman[83286]: 2025-11-27 08:29:57.09294407 +0000 UTC m=+0.188905205 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, release=1761123044, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, vcs-type=git, url=https://www.redhat.com, distribution-scope=public, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:29:57 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:29:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:29:59 localhost systemd[1]: tmp-crun.oPqzqn.mount: Deactivated successfully. Nov 27 03:29:59 localhost podman[83332]: 2025-11-27 08:29:59.007699466 +0000 UTC m=+0.088441151 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, architecture=x86_64, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, container_name=nova_compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container) Nov 27 03:29:59 localhost podman[83332]: 2025-11-27 08:29:59.061655653 +0000 UTC m=+0.142397338 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, container_name=nova_compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, url=https://www.redhat.com, config_id=tripleo_step5, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-nova-compute-container) Nov 27 03:29:59 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:30:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:30:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:30:06 localhost podman[83437]: 2025-11-27 08:30:06.997140894 +0000 UTC m=+0.092631073 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, distribution-scope=public, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, container_name=iscsid, com.redhat.component=openstack-iscsid-container, release=1761123044) Nov 27 03:30:07 localhost podman[83437]: 2025-11-27 08:30:07.040961908 +0000 UTC m=+0.136452117 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, batch=17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., version=17.1.12, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.buildah.version=1.41.4, container_name=iscsid, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container) Nov 27 03:30:07 localhost podman[83436]: 2025-11-27 08:30:07.048562413 +0000 UTC m=+0.144192896 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, release=1761123044, distribution-scope=public, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.openshift.expose-services=, config_id=tripleo_step3, tcib_managed=true, vcs-type=git, com.redhat.component=openstack-collectd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com) Nov 27 03:30:07 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:30:07 localhost podman[83436]: 2025-11-27 08:30:07.063411944 +0000 UTC m=+0.159042407 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.buildah.version=1.41.4, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step3, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, container_name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., architecture=x86_64, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd) Nov 27 03:30:07 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:30:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:30:17 localhost podman[83473]: 2025-11-27 08:30:17.991823762 +0000 UTC m=+0.085741587 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, container_name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-qdrouterd) Nov 27 03:30:18 localhost podman[83473]: 2025-11-27 08:30:18.188937348 +0000 UTC m=+0.282855163 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, url=https://www.redhat.com, config_id=tripleo_step1, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, build-date=2025-11-18T22:49:46Z, version=17.1.12, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:30:18 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:30:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:30:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:30:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:30:21 localhost podman[83500]: 2025-11-27 08:30:21.995750874 +0000 UTC m=+0.088894153 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, vendor=Red Hat, Inc., config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, managed_by=tripleo_ansible, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, vcs-type=git, url=https://www.redhat.com) Nov 27 03:30:22 localhost podman[83501]: 2025-11-27 08:30:22.051115588 +0000 UTC m=+0.138443880 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, build-date=2025-11-18T22:49:32Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-cron-container, distribution-scope=public, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, tcib_managed=true, name=rhosp17/openstack-cron) Nov 27 03:30:22 localhost podman[83501]: 2025-11-27 08:30:22.090197944 +0000 UTC m=+0.177526236 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., release=1761123044, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-type=git, config_id=tripleo_step4, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:30:22 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:30:22 localhost podman[83502]: 2025-11-27 08:30:22.11409291 +0000 UTC m=+0.200004154 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-type=git, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., container_name=ceilometer_agent_compute, version=17.1.12) Nov 27 03:30:22 localhost podman[83500]: 2025-11-27 08:30:22.13072806 +0000 UTC m=+0.223871329 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12) Nov 27 03:30:22 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:30:22 localhost podman[83502]: 2025-11-27 08:30:22.148112299 +0000 UTC m=+0.234023493 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, tcib_managed=true, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, release=1761123044, io.openshift.expose-services=, container_name=ceilometer_agent_compute) Nov 27 03:30:22 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:30:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:30:24 localhost systemd[1]: tmp-crun.lKWR6b.mount: Deactivated successfully. Nov 27 03:30:24 localhost podman[83571]: 2025-11-27 08:30:24.996662195 +0000 UTC m=+0.091512534 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, architecture=x86_64, config_id=tripleo_step4, container_name=nova_migration_target, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git) Nov 27 03:30:25 localhost podman[83571]: 2025-11-27 08:30:25.373725532 +0000 UTC m=+0.468575881 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, tcib_managed=true, batch=17.1_20251118.1, io.openshift.expose-services=, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, distribution-scope=public, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:30:25 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:30:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:30:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:30:27 localhost systemd[1]: tmp-crun.BnmUDQ.mount: Deactivated successfully. Nov 27 03:30:27 localhost podman[83593]: 2025-11-27 08:30:27.998805522 +0000 UTC m=+0.092469219 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_id=tripleo_step4, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, managed_by=tripleo_ansible, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, tcib_managed=true, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, io.openshift.expose-services=, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, distribution-scope=public, build-date=2025-11-18T23:34:05Z, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:30:28 localhost podman[83593]: 2025-11-27 08:30:28.030503059 +0000 UTC m=+0.124166746 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, architecture=x86_64, vcs-type=git, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, config_id=tripleo_step4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, distribution-scope=public, managed_by=tripleo_ansible, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:30:28 localhost podman[83592]: 2025-11-27 08:30:28.045876984 +0000 UTC m=+0.139836560 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, architecture=x86_64, container_name=ovn_metadata_agent, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn) Nov 27 03:30:28 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:30:28 localhost podman[83592]: 2025-11-27 08:30:28.114914579 +0000 UTC m=+0.208874125 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, version=17.1.12, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, io.buildah.version=1.41.4, tcib_managed=true, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:30:28 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:30:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:30:30 localhost podman[83639]: 2025-11-27 08:30:30.0024829 +0000 UTC m=+0.091476983 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, tcib_managed=true, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.openshift.expose-services=, managed_by=tripleo_ansible, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_compute, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container) Nov 27 03:30:30 localhost podman[83639]: 2025-11-27 08:30:30.031031711 +0000 UTC m=+0.120025754 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vcs-type=git, distribution-scope=public, managed_by=tripleo_ansible, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, architecture=x86_64, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:30:30 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:30:32 localhost systemd[1]: session-34.scope: Deactivated successfully. Nov 27 03:30:32 localhost systemd[1]: session-34.scope: Consumed 13.223s CPU time. Nov 27 03:30:32 localhost systemd-logind[761]: Session 34 logged out. Waiting for processes to exit. Nov 27 03:30:32 localhost systemd-logind[761]: Removed session 34. Nov 27 03:30:37 localhost sshd[83665]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:30:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:30:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:30:37 localhost podman[83667]: 2025-11-27 08:30:37.840422245 +0000 UTC m=+0.089670694 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, io.buildah.version=1.41.4, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, version=17.1.12, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container) Nov 27 03:30:37 localhost systemd-logind[761]: New session 35 of user zuul. Nov 27 03:30:37 localhost systemd[1]: Started Session 35 of User zuul. Nov 27 03:30:37 localhost podman[83667]: 2025-11-27 08:30:37.88577231 +0000 UTC m=+0.135020739 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, tcib_managed=true, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., container_name=collectd, distribution-scope=public, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, release=1761123044, maintainer=OpenStack TripleO Team) Nov 27 03:30:37 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:30:37 localhost podman[83668]: 2025-11-27 08:30:37.888102753 +0000 UTC m=+0.132449720 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, architecture=x86_64, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, vendor=Red Hat, Inc., managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, version=17.1.12, tcib_managed=true) Nov 27 03:30:37 localhost podman[83668]: 2025-11-27 08:30:37.967774365 +0000 UTC m=+0.212121332 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, name=rhosp17/openstack-iscsid, release=1761123044, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.buildah.version=1.41.4, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:30:37 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:30:38 localhost python3[83724]: ansible-ansible.legacy.command Invoked with _raw_params=subscription-manager repos --disable rhceph-7-tools-for-rhel-9-x86_64-rpms _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 03:30:42 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 03:30:43 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 03:30:44 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:30:44 localhost recover_tripleo_nova_virtqemud[83856]: 63639 Nov 27 03:30:44 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:30:44 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:30:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:30:48 localhost podman[83914]: 2025-11-27 08:30:48.992853174 +0000 UTC m=+0.084597258 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-type=git, version=17.1.12, distribution-scope=public, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:30:49 localhost podman[83914]: 2025-11-27 08:30:49.235296294 +0000 UTC m=+0.327040388 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, release=1761123044, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com) Nov 27 03:30:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:30:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:30:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:30:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:30:53 localhost systemd[1]: tmp-crun.Jryxag.mount: Deactivated successfully. Nov 27 03:30:53 localhost podman[83965]: 2025-11-27 08:30:53.019246503 +0000 UTC m=+0.104681659 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, tcib_managed=true, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64) Nov 27 03:30:53 localhost podman[83967]: 2025-11-27 08:30:53.064675101 +0000 UTC m=+0.143212331 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:30:53 localhost podman[83965]: 2025-11-27 08:30:53.074441045 +0000 UTC m=+0.159876181 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:30:53 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:30:53 localhost podman[83967]: 2025-11-27 08:30:53.125164215 +0000 UTC m=+0.203701525 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, architecture=x86_64, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.expose-services=, batch=17.1_20251118.1, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:30:53 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:30:53 localhost podman[83966]: 2025-11-27 08:30:53.126808939 +0000 UTC m=+0.206990054 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, vcs-type=git, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, version=17.1.12, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, distribution-scope=public, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc.) Nov 27 03:30:53 localhost podman[83966]: 2025-11-27 08:30:53.206789171 +0000 UTC m=+0.286970236 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, managed_by=tripleo_ansible, url=https://www.redhat.com, vcs-type=git, distribution-scope=public, name=rhosp17/openstack-cron, architecture=x86_64, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:32Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, vendor=Red Hat, Inc.) Nov 27 03:30:53 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:30:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:30:55 localhost podman[84061]: 2025-11-27 08:30:55.984655417 +0000 UTC m=+0.079199142 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, release=1761123044, container_name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, io.openshift.expose-services=, vendor=Red Hat, Inc., managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, version=17.1.12, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:30:56 localhost podman[84061]: 2025-11-27 08:30:56.363265757 +0000 UTC m=+0.457809522 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, tcib_managed=true) Nov 27 03:30:56 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:30:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:30:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:30:58 localhost systemd[1]: tmp-crun.Lspqam.mount: Deactivated successfully. Nov 27 03:30:58 localhost podman[84085]: 2025-11-27 08:30:58.998766025 +0000 UTC m=+0.097266998 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, tcib_managed=true, architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, release=1761123044, io.openshift.expose-services=) Nov 27 03:30:59 localhost podman[84086]: 2025-11-27 08:30:59.038567291 +0000 UTC m=+0.133064247 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, version=17.1.12, container_name=ovn_controller, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vendor=Red Hat, Inc., io.openshift.expose-services=, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64) Nov 27 03:30:59 localhost podman[84085]: 2025-11-27 08:30:59.074050379 +0000 UTC m=+0.172551312 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.buildah.version=1.41.4, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, managed_by=tripleo_ansible, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, tcib_managed=true, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:30:59 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:30:59 localhost podman[84086]: 2025-11-27 08:30:59.09110821 +0000 UTC m=+0.185605196 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, distribution-scope=public, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, tcib_managed=true, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1) Nov 27 03:30:59 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:31:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:31:00 localhost podman[84147]: 2025-11-27 08:31:00.845158574 +0000 UTC m=+0.086244782 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, release=1761123044, managed_by=tripleo_ansible, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:31:00 localhost podman[84147]: 2025-11-27 08:31:00.873615793 +0000 UTC m=+0.114701981 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, release=1761123044, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, vcs-type=git, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:31:00 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:31:01 localhost podman[84258]: 2025-11-27 08:31:01.711109261 +0000 UTC m=+0.097430503 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, com.redhat.component=rhceph-container, vcs-type=git, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, distribution-scope=public, description=Red Hat Ceph Storage 7, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 03:31:01 localhost podman[84258]: 2025-11-27 08:31:01.820668151 +0000 UTC m=+0.206989433 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , RELEASE=main, name=rhceph, GIT_BRANCH=main, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_CLEAN=True, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, vcs-type=git, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph) Nov 27 03:31:02 localhost python3[84326]: ansible-ansible.builtin.slurp Invoked with path=/home/zuul/ansible_hostname src=/home/zuul/ansible_hostname Nov 27 03:31:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:31:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:31:08 localhost podman[84415]: 2025-11-27 08:31:08.997176106 +0000 UTC m=+0.091004201 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., release=1761123044, container_name=collectd, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, distribution-scope=public) Nov 27 03:31:09 localhost podman[84416]: 2025-11-27 08:31:09.042955932 +0000 UTC m=+0.137033803 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, release=1761123044, build-date=2025-11-18T23:44:13Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, container_name=iscsid) Nov 27 03:31:09 localhost podman[84416]: 2025-11-27 08:31:09.058010679 +0000 UTC m=+0.152088550 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, container_name=iscsid, architecture=x86_64, vcs-type=git, name=rhosp17/openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, tcib_managed=true) Nov 27 03:31:09 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:31:09 localhost podman[84415]: 2025-11-27 08:31:09.112324557 +0000 UTC m=+0.206152642 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, container_name=collectd, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, version=17.1.12, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public) Nov 27 03:31:09 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:31:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:31:19 localhost podman[84455]: 2025-11-27 08:31:19.992275543 +0000 UTC m=+0.086481157 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., managed_by=tripleo_ansible, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, version=17.1.12) Nov 27 03:31:20 localhost podman[84455]: 2025-11-27 08:31:20.189007119 +0000 UTC m=+0.283212713 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, container_name=metrics_qdr, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, url=https://www.redhat.com, release=1761123044, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd) Nov 27 03:31:20 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:31:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:31:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:31:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:31:23 localhost podman[84485]: 2025-11-27 08:31:23.990218885 +0000 UTC m=+0.081964646 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, build-date=2025-11-18T22:49:32Z, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-cron-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, version=17.1.12, vcs-type=git, io.openshift.expose-services=, config_id=tripleo_step4, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true) Nov 27 03:31:24 localhost podman[84485]: 2025-11-27 08:31:24.00187236 +0000 UTC m=+0.093618131 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-type=git, container_name=logrotate_crond, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, io.openshift.expose-services=, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, version=17.1.12, distribution-scope=public, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container) Nov 27 03:31:24 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:31:24 localhost podman[84484]: 2025-11-27 08:31:24.097314719 +0000 UTC m=+0.190046656 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, distribution-scope=public, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, version=17.1.12, release=1761123044, name=rhosp17/openstack-ceilometer-ipmi, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4) Nov 27 03:31:24 localhost podman[84486]: 2025-11-27 08:31:24.151879053 +0000 UTC m=+0.238015282 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, url=https://www.redhat.com, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, tcib_managed=true, com.redhat.component=openstack-ceilometer-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, io.openshift.expose-services=, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:31:24 localhost podman[84484]: 2025-11-27 08:31:24.156040035 +0000 UTC m=+0.248771982 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-ipmi-container, version=17.1.12, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, url=https://www.redhat.com) Nov 27 03:31:24 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:31:24 localhost podman[84486]: 2025-11-27 08:31:24.187100455 +0000 UTC m=+0.273236664 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, batch=17.1_20251118.1, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, tcib_managed=true, build-date=2025-11-19T00:11:48Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public) Nov 27 03:31:24 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:31:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:31:27 localhost podman[84557]: 2025-11-27 08:31:26.982563596 +0000 UTC m=+0.080606400 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, maintainer=OpenStack TripleO Team, version=17.1.12, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, release=1761123044, io.openshift.expose-services=, url=https://www.redhat.com, vcs-type=git, config_id=tripleo_step4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container) Nov 27 03:31:27 localhost podman[84557]: 2025-11-27 08:31:27.34730689 +0000 UTC m=+0.445349684 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_migration_target, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_id=tripleo_step4, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 03:31:27 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:31:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:31:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:31:29 localhost podman[84580]: 2025-11-27 08:31:29.99169589 +0000 UTC m=+0.090766814 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, tcib_managed=true, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, release=1761123044) Nov 27 03:31:30 localhost systemd[1]: tmp-crun.Gq3v90.mount: Deactivated successfully. Nov 27 03:31:30 localhost podman[84580]: 2025-11-27 08:31:30.048970777 +0000 UTC m=+0.148041741 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, architecture=x86_64, managed_by=tripleo_ansible, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:31:30 localhost podman[84581]: 2025-11-27 08:31:30.047128708 +0000 UTC m=+0.141616018 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, version=17.1.12, tcib_managed=true, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:31:30 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:31:30 localhost podman[84581]: 2025-11-27 08:31:30.133943133 +0000 UTC m=+0.228430463 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, build-date=2025-11-18T23:34:05Z, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, maintainer=OpenStack TripleO Team, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, distribution-scope=public, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, batch=17.1_20251118.1, vcs-type=git, io.openshift.expose-services=, architecture=x86_64, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:31:30 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:31:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:31:31 localhost podman[84627]: 2025-11-27 08:31:31.986366063 +0000 UTC m=+0.084228016 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, distribution-scope=public, io.buildah.version=1.41.4, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, vcs-type=git, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:31:32 localhost podman[84627]: 2025-11-27 08:31:32.018010798 +0000 UTC m=+0.115872751 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, batch=17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, release=1761123044, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute) Nov 27 03:31:32 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:31:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:31:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:31:39 localhost podman[84654]: 2025-11-27 08:31:39.986399499 +0000 UTC m=+0.081026960 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, maintainer=OpenStack TripleO Team, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, batch=17.1_20251118.1, name=rhosp17/openstack-iscsid, vcs-type=git, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=) Nov 27 03:31:39 localhost podman[84654]: 2025-11-27 08:31:39.995129445 +0000 UTC m=+0.089756906 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, release=1761123044, vcs-type=git, io.buildah.version=1.41.4, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, name=rhosp17/openstack-iscsid, io.openshift.expose-services=) Nov 27 03:31:40 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:31:40 localhost systemd[1]: tmp-crun.HMumVS.mount: Deactivated successfully. Nov 27 03:31:40 localhost podman[84653]: 2025-11-27 08:31:40.043472621 +0000 UTC m=+0.140889367 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, com.redhat.component=openstack-collectd-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, config_id=tripleo_step3, vcs-type=git, distribution-scope=public, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, container_name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:31:40 localhost podman[84653]: 2025-11-27 08:31:40.055942117 +0000 UTC m=+0.153358903 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, name=rhosp17/openstack-collectd, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, config_id=tripleo_step3) Nov 27 03:31:40 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:31:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:31:51 localhost podman[84690]: 2025-11-27 08:31:51.033187474 +0000 UTC m=+0.124392622 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, tcib_managed=true, vcs-type=git, release=1761123044, version=17.1.12, distribution-scope=public, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd) Nov 27 03:31:51 localhost podman[84690]: 2025-11-27 08:31:51.26769504 +0000 UTC m=+0.358900168 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, url=https://www.redhat.com, vcs-type=git, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., container_name=metrics_qdr, tcib_managed=true, release=1761123044, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, version=17.1.12) Nov 27 03:31:51 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:31:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:31:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:31:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:31:54 localhost podman[84764]: 2025-11-27 08:31:54.992257135 +0000 UTC m=+0.085375198 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vendor=Red Hat, Inc., release=1761123044, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, managed_by=tripleo_ansible, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git) Nov 27 03:31:55 localhost podman[84764]: 2025-11-27 08:31:55.024966069 +0000 UTC m=+0.118084142 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.expose-services=, release=1761123044, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, version=17.1.12, vendor=Red Hat, Inc., config_id=tripleo_step4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:31:55 localhost podman[84766]: 2025-11-27 08:31:55.05350179 +0000 UTC m=+0.140439486 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_id=tripleo_step4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, vcs-type=git, tcib_managed=true, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, io.buildah.version=1.41.4) Nov 27 03:31:55 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:31:55 localhost podman[84766]: 2025-11-27 08:31:55.091298451 +0000 UTC m=+0.178236187 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, tcib_managed=true, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, distribution-scope=public, batch=17.1_20251118.1, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, io.openshift.expose-services=, config_id=tripleo_step4) Nov 27 03:31:55 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:31:55 localhost podman[84765]: 2025-11-27 08:31:55.156871373 +0000 UTC m=+0.247572660 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-cron-container, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, distribution-scope=public, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64) Nov 27 03:31:55 localhost podman[84765]: 2025-11-27 08:31:55.170885881 +0000 UTC m=+0.261587158 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, tcib_managed=true, config_id=tripleo_step4, container_name=logrotate_crond, vendor=Red Hat, Inc., managed_by=tripleo_ansible, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:31:55 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:31:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:31:57 localhost podman[84837]: 2025-11-27 08:31:57.992698435 +0000 UTC m=+0.085763418 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, batch=17.1_20251118.1, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, release=1761123044, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible) Nov 27 03:31:58 localhost podman[84837]: 2025-11-27 08:31:58.372015914 +0000 UTC m=+0.465080927 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-type=git, version=17.1.12, container_name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z) Nov 27 03:31:58 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:32:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:32:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:32:00 localhost systemd[1]: tmp-crun.ip6ko3.mount: Deactivated successfully. Nov 27 03:32:00 localhost podman[84859]: 2025-11-27 08:32:00.996113675 +0000 UTC m=+0.095524532 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, tcib_managed=true, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12) Nov 27 03:32:01 localhost podman[84859]: 2025-11-27 08:32:01.087498435 +0000 UTC m=+0.186909242 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, config_id=tripleo_step4, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, version=17.1.12, architecture=x86_64, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, batch=17.1_20251118.1) Nov 27 03:32:01 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:32:01 localhost podman[84860]: 2025-11-27 08:32:01.16761069 +0000 UTC m=+0.264513549 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, container_name=ovn_controller, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.expose-services=, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git) Nov 27 03:32:01 localhost podman[84860]: 2025-11-27 08:32:01.217501137 +0000 UTC m=+0.314403976 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, version=17.1.12, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:32:01 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:32:02 localhost systemd[1]: session-35.scope: Deactivated successfully. Nov 27 03:32:02 localhost systemd[1]: session-35.scope: Consumed 6.509s CPU time. Nov 27 03:32:02 localhost systemd-logind[761]: Session 35 logged out. Waiting for processes to exit. Nov 27 03:32:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:32:02 localhost systemd-logind[761]: Removed session 35. Nov 27 03:32:02 localhost podman[84908]: 2025-11-27 08:32:02.408133907 +0000 UTC m=+0.083191738 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, tcib_managed=true, vendor=Red Hat, Inc., container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:32:02 localhost podman[84908]: 2025-11-27 08:32:02.439244538 +0000 UTC m=+0.114302369 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, distribution-scope=public, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:32:02 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:32:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:32:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:32:11 localhost podman[85010]: 2025-11-27 08:32:11.000085266 +0000 UTC m=+0.093175519 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, build-date=2025-11-18T22:51:28Z, container_name=collectd) Nov 27 03:32:11 localhost podman[85011]: 2025-11-27 08:32:11.053955072 +0000 UTC m=+0.146450929 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, vcs-type=git, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044) Nov 27 03:32:11 localhost podman[85010]: 2025-11-27 08:32:11.067773745 +0000 UTC m=+0.160863978 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, vcs-type=git, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., version=17.1.12, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}) Nov 27 03:32:11 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:32:11 localhost podman[85011]: 2025-11-27 08:32:11.09092089 +0000 UTC m=+0.183416797 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, release=1761123044, distribution-scope=public, version=17.1.12, batch=17.1_20251118.1, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-type=git, io.openshift.expose-services=, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team) Nov 27 03:32:11 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:32:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:32:22 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:32:22 localhost recover_tripleo_nova_virtqemud[85050]: 63639 Nov 27 03:32:22 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:32:22 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:32:22 localhost podman[85048]: 2025-11-27 08:32:22.211650707 +0000 UTC m=+0.103890098 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, batch=17.1_20251118.1, tcib_managed=true, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.openshift.expose-services=, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z) Nov 27 03:32:22 localhost podman[85048]: 2025-11-27 08:32:22.42906979 +0000 UTC m=+0.321309241 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, tcib_managed=true, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, url=https://www.redhat.com, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, version=17.1.12) Nov 27 03:32:22 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:32:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:32:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:32:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:32:26 localhost systemd[1]: tmp-crun.Bqg9Us.mount: Deactivated successfully. Nov 27 03:32:26 localhost podman[85079]: 2025-11-27 08:32:26.010614151 +0000 UTC m=+0.104475924 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, url=https://www.redhat.com, managed_by=tripleo_ansible, distribution-scope=public, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:32:26 localhost podman[85079]: 2025-11-27 08:32:26.040103078 +0000 UTC m=+0.133964851 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., managed_by=tripleo_ansible, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:32:26 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:32:26 localhost podman[85081]: 2025-11-27 08:32:26.097478818 +0000 UTC m=+0.185192294 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, distribution-scope=public, version=17.1.12, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, release=1761123044, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:32:26 localhost podman[85080]: 2025-11-27 08:32:26.150625594 +0000 UTC m=+0.241319221 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, version=17.1.12, name=rhosp17/openstack-cron, config_id=tripleo_step4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:32:26 localhost podman[85080]: 2025-11-27 08:32:26.187900441 +0000 UTC m=+0.278594068 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com) Nov 27 03:32:26 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:32:26 localhost podman[85081]: 2025-11-27 08:32:26.203440781 +0000 UTC m=+0.291154257 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., distribution-scope=public, version=17.1.12, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64) Nov 27 03:32:26 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:32:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:32:28 localhost podman[85153]: 2025-11-27 08:32:28.981387568 +0000 UTC m=+0.076639831 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, release=1761123044, container_name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.buildah.version=1.41.4, version=17.1.12) Nov 27 03:32:29 localhost podman[85153]: 2025-11-27 08:32:29.383467453 +0000 UTC m=+0.478719716 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, release=1761123044, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:32:29 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:32:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:32:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:32:31 localhost podman[85175]: 2025-11-27 08:32:31.989087074 +0000 UTC m=+0.085766808 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, tcib_managed=true, architecture=x86_64, config_id=tripleo_step4) Nov 27 03:32:32 localhost podman[85176]: 2025-11-27 08:32:32.039285351 +0000 UTC m=+0.133066067 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vendor=Red Hat, Inc., tcib_managed=true, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, io.openshift.expose-services=, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, container_name=ovn_controller, version=17.1.12) Nov 27 03:32:32 localhost podman[85175]: 2025-11-27 08:32:32.061899082 +0000 UTC m=+0.158578876 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vendor=Red Hat, Inc., url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, batch=17.1_20251118.1, container_name=ovn_metadata_agent, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:32:32 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:32:32 localhost podman[85176]: 2025-11-27 08:32:32.093877315 +0000 UTC m=+0.187658031 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, architecture=x86_64, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:32:32 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:32:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:32:32 localhost systemd[1]: tmp-crun.GpsFqO.mount: Deactivated successfully. Nov 27 03:32:32 localhost podman[85221]: 2025-11-27 08:32:32.993931575 +0000 UTC m=+0.093057746 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, batch=17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., distribution-scope=public) Nov 27 03:32:33 localhost podman[85221]: 2025-11-27 08:32:33.049957679 +0000 UTC m=+0.149083820 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.expose-services=, batch=17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step5, url=https://www.redhat.com, version=17.1.12, release=1761123044, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc.) Nov 27 03:32:33 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:32:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:32:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:32:41 localhost podman[85248]: 2025-11-27 08:32:41.987692009 +0000 UTC m=+0.082478599 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, release=1761123044, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, tcib_managed=true, io.openshift.expose-services=) Nov 27 03:32:42 localhost podman[85248]: 2025-11-27 08:32:41.99996199 +0000 UTC m=+0.094748550 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, build-date=2025-11-18T22:51:28Z, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, name=rhosp17/openstack-collectd, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=) Nov 27 03:32:42 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:32:42 localhost podman[85249]: 2025-11-27 08:32:42.090742804 +0000 UTC m=+0.180753306 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, tcib_managed=true, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public) Nov 27 03:32:42 localhost podman[85249]: 2025-11-27 08:32:42.125207495 +0000 UTC m=+0.215218007 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, architecture=x86_64, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-18T23:44:13Z, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:32:42 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:32:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:32:52 localhost podman[85287]: 2025-11-27 08:32:52.994905526 +0000 UTC m=+0.090536147 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, architecture=x86_64, config_id=tripleo_step1, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, io.openshift.expose-services=, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr) Nov 27 03:32:53 localhost podman[85287]: 2025-11-27 08:32:53.21043455 +0000 UTC m=+0.306065101 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, tcib_managed=true, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, version=17.1.12, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:32:53 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:32:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:32:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:32:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:32:56 localhost podman[85362]: 2025-11-27 08:32:56.990884265 +0000 UTC m=+0.083630681 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, vcs-type=git, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, version=17.1.12, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z) Nov 27 03:32:57 localhost podman[85363]: 2025-11-27 08:32:57.054256786 +0000 UTC m=+0.141273657 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4) Nov 27 03:32:57 localhost podman[85363]: 2025-11-27 08:32:57.085878091 +0000 UTC m=+0.172894982 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, version=17.1.12) Nov 27 03:32:57 localhost podman[85361]: 2025-11-27 08:32:57.095002128 +0000 UTC m=+0.187237910 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, version=17.1.12, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, url=https://www.redhat.com, architecture=x86_64, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:32:57 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:32:57 localhost podman[85361]: 2025-11-27 08:32:57.123006924 +0000 UTC m=+0.215242666 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-type=git, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1) Nov 27 03:32:57 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:32:57 localhost podman[85362]: 2025-11-27 08:32:57.174908896 +0000 UTC m=+0.267655302 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, config_id=tripleo_step4, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, vcs-type=git, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, distribution-scope=public, tcib_managed=true, release=1761123044, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container) Nov 27 03:32:57 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:32:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:33:00 localhost podman[85433]: 2025-11-27 08:32:59.999120775 +0000 UTC m=+0.087453784 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64) Nov 27 03:33:00 localhost podman[85433]: 2025-11-27 08:33:00.430356116 +0000 UTC m=+0.518689135 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, version=17.1.12, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, vcs-type=git) Nov 27 03:33:00 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:33:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:33:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:33:02 localhost podman[85479]: 2025-11-27 08:33:02.992300258 +0000 UTC m=+0.079315424 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public) Nov 27 03:33:03 localhost podman[85479]: 2025-11-27 08:33:03.040970372 +0000 UTC m=+0.127985538 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, container_name=ovn_metadata_agent, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:33:03 localhost systemd[1]: tmp-crun.qk0cqh.mount: Deactivated successfully. Nov 27 03:33:03 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:33:03 localhost podman[85480]: 2025-11-27 08:33:03.063154082 +0000 UTC m=+0.145949774 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, release=1761123044, architecture=x86_64, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, vcs-type=git, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vendor=Red Hat, Inc., distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:33:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:33:03 localhost podman[85480]: 2025-11-27 08:33:03.115029404 +0000 UTC m=+0.197825076 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, release=1761123044, vcs-type=git, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., version=17.1.12, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:33:03 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:33:03 localhost podman[85525]: 2025-11-27 08:33:03.190794891 +0000 UTC m=+0.088694858 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, vcs-type=git, vendor=Red Hat, Inc., release=1761123044, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, architecture=x86_64, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12) Nov 27 03:33:03 localhost podman[85525]: 2025-11-27 08:33:03.25513156 +0000 UTC m=+0.153031527 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, distribution-scope=public, vendor=Red Hat, Inc., container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_id=tripleo_step5, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container) Nov 27 03:33:03 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:33:09 localhost systemd-logind[761]: Existing logind session ID 28 used by new audit session, ignoring. Nov 27 03:33:09 localhost systemd[1]: Created slice User Slice of UID 0. Nov 27 03:33:09 localhost systemd[1]: Starting User Runtime Directory /run/user/0... Nov 27 03:33:09 localhost systemd[1]: Finished User Runtime Directory /run/user/0. Nov 27 03:33:09 localhost systemd[1]: Starting User Manager for UID 0... Nov 27 03:33:09 localhost systemd[85983]: Queued start job for default target Main User Target. Nov 27 03:33:09 localhost systemd[85983]: Created slice User Application Slice. Nov 27 03:33:09 localhost systemd[85983]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). Nov 27 03:33:09 localhost systemd[85983]: Started Daily Cleanup of User's Temporary Directories. Nov 27 03:33:09 localhost systemd[85983]: Reached target Paths. Nov 27 03:33:09 localhost systemd[85983]: Reached target Timers. Nov 27 03:33:09 localhost systemd[85983]: Starting D-Bus User Message Bus Socket... Nov 27 03:33:09 localhost systemd[85983]: Starting Create User's Volatile Files and Directories... Nov 27 03:33:09 localhost systemd[85983]: Listening on D-Bus User Message Bus Socket. Nov 27 03:33:09 localhost systemd[85983]: Reached target Sockets. Nov 27 03:33:09 localhost systemd[85983]: Finished Create User's Volatile Files and Directories. Nov 27 03:33:09 localhost systemd[85983]: Reached target Basic System. Nov 27 03:33:09 localhost systemd[85983]: Reached target Main User Target. Nov 27 03:33:09 localhost systemd[85983]: Startup finished in 155ms. Nov 27 03:33:09 localhost systemd[1]: Started User Manager for UID 0. Nov 27 03:33:09 localhost systemd[1]: Started Session c11 of User root. Nov 27 03:33:10 localhost kernel: tun: Universal TUN/TAP device driver, 1.6 Nov 27 03:33:10 localhost kernel: device tapa2718872-3b entered promiscuous mode Nov 27 03:33:10 localhost NetworkManager[5971]: [1764232390.8744] manager: (tapa2718872-3b): new Tun device (/org/freedesktop/NetworkManager/Devices/13) Nov 27 03:33:10 localhost systemd-udevd[86018]: Network interface NamePolicy= disabled on kernel command line. Nov 27 03:33:10 localhost NetworkManager[5971]: [1764232390.9026] device (tapa2718872-3b): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 03:33:10 localhost systemd[1]: Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 27 03:33:10 localhost NetworkManager[5971]: [1764232390.9037] device (tapa2718872-3b): state change: unavailable -> disconnected (reason 'none', sys-iface-state: 'external') Nov 27 03:33:10 localhost systemd[1]: Starting Virtual Machine and Container Registration Service... Nov 27 03:33:10 localhost systemd[1]: Started Virtual Machine and Container Registration Service. Nov 27 03:33:10 localhost systemd-machined[86026]: New machine qemu-1-instance-00000002. Nov 27 03:33:10 localhost systemd[1]: Started Virtual Machine qemu-1-instance-00000002. Nov 27 03:33:11 localhost NetworkManager[5971]: [1764232391.1549] manager: (tap3d77aad9-80): new Veth device (/org/freedesktop/NetworkManager/Devices/14) Nov 27 03:33:11 localhost systemd-udevd[86017]: Network interface NamePolicy= disabled on kernel command line. Nov 27 03:33:11 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap3d77aad9-81: link becomes ready Nov 27 03:33:11 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap3d77aad9-80: link becomes ready Nov 27 03:33:11 localhost NetworkManager[5971]: [1764232391.2039] device (tap3d77aad9-80): carrier: link connected Nov 27 03:33:11 localhost kernel: device tap3d77aad9-80 entered promiscuous mode Nov 27 03:33:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:33:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:33:12 localhost systemd[1]: Starting SETroubleshoot daemon for processing new SELinux denial logs... Nov 27 03:33:12 localhost systemd[1]: tmp-crun.ecl9Go.mount: Deactivated successfully. Nov 27 03:33:12 localhost podman[86123]: 2025-11-27 08:33:12.755125711 +0000 UTC m=+0.114218327 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_id=tripleo_step3, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-iscsid-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:33:12 localhost podman[86123]: 2025-11-27 08:33:12.761724449 +0000 UTC m=+0.120817065 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, maintainer=OpenStack TripleO Team, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, version=17.1.12, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, tcib_managed=true) Nov 27 03:33:12 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:33:12 localhost podman[86122]: 2025-11-27 08:33:12.846303265 +0000 UTC m=+0.207482797 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, vcs-type=git, container_name=collectd, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:33:12 localhost podman[86122]: 2025-11-27 08:33:12.887831057 +0000 UTC m=+0.249010539 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4, managed_by=tripleo_ansible, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, distribution-scope=public, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:33:12 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:33:12 localhost systemd[1]: Started SETroubleshoot daemon for processing new SELinux denial logs. Nov 27 03:33:13 localhost podman[86187]: 2025-11-27 08:33:13.038361704 +0000 UTC m=+0.067028312 container create 8edf4b618175d9655584ba4583d12318595037d7baa99238850bc413023b35da (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, tcib_managed=true, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, vendor=Red Hat, Inc., build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:33:13 localhost systemd[1]: Started libpod-conmon-8edf4b618175d9655584ba4583d12318595037d7baa99238850bc413023b35da.scope. Nov 27 03:33:13 localhost podman[86187]: 2025-11-27 08:33:13.00413573 +0000 UTC m=+0.032802338 image pull registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1 Nov 27 03:33:13 localhost systemd[1]: Started libcrun container. Nov 27 03:33:13 localhost systemd[1]: Created slice Slice /system/dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged. Nov 27 03:33:13 localhost systemd[1]: Started dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@0.service. Nov 27 03:33:13 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1e52d522d7c1eea0db09ac0f1708dae78ab9f93d17e88b96d5bc45b47aceab37/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 03:33:13 localhost podman[86187]: 2025-11-27 08:33:13.144849811 +0000 UTC m=+0.173516439 container init 8edf4b618175d9655584ba4583d12318595037d7baa99238850bc413023b35da (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, io.openshift.expose-services=, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, release=1761123044, tcib_managed=true, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:33:13 localhost podman[86187]: 2025-11-27 08:33:13.153610728 +0000 UTC m=+0.182277356 container start 8edf4b618175d9655584ba4583d12318595037d7baa99238850bc413023b35da (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:33:14 localhost setroubleshoot[86124]: SELinux is preventing /usr/libexec/qemu-kvm from read access on the file max_map_count. For complete SELinux messages run: sealert -l 16fdbab6-fdc7-40ae-9244-67863b31fd5b Nov 27 03:33:14 localhost setroubleshoot[86124]: SELinux is preventing /usr/libexec/qemu-kvm from read access on the file max_map_count.#012#012***** Plugin qemu_file_image (98.8 confidence) suggests *******************#012#012If max_map_count is a virtualization target#012Then you need to change the label on max_map_count'#012Do#012# semanage fcontext -a -t virt_image_t 'max_map_count'#012# restorecon -v 'max_map_count'#012#012***** Plugin catchall (2.13 confidence) suggests **************************#012#012If you believe that qemu-kvm should be allowed read access on the max_map_count file by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'qemu-kvm' --raw | audit2allow -M my-qemukvm#012# semodule -X 300 -i my-qemukvm.pp#012 Nov 27 03:33:21 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #43. Immutable memtables: 0. Nov 27 03:33:23 localhost systemd[1]: dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@0.service: Deactivated successfully. Nov 27 03:33:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:33:23 localhost podman[86224]: 2025-11-27 08:33:23.590645099 +0000 UTC m=+0.091729729 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, build-date=2025-11-18T22:49:46Z, version=17.1.12, container_name=metrics_qdr, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd) Nov 27 03:33:23 localhost podman[86224]: 2025-11-27 08:33:23.777005814 +0000 UTC m=+0.278090444 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, architecture=x86_64, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:33:23 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:33:24 localhost systemd[1]: setroubleshootd.service: Deactivated successfully. Nov 27 03:33:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:33:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:33:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:33:27 localhost systemd[1]: tmp-crun.4LnbE5.mount: Deactivated successfully. Nov 27 03:33:28 localhost podman[86255]: 2025-11-27 08:33:28.000241464 +0000 UTC m=+0.094649449 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, vcs-type=git, io.openshift.expose-services=, distribution-scope=public, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044) Nov 27 03:33:28 localhost podman[86256]: 2025-11-27 08:33:28.053044201 +0000 UTC m=+0.141755492 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, name=rhosp17/openstack-cron, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_id=tripleo_step4, io.openshift.expose-services=, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team) Nov 27 03:33:28 localhost podman[86256]: 2025-11-27 08:33:28.091837969 +0000 UTC m=+0.180549260 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, distribution-scope=public, tcib_managed=true, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step4, release=1761123044, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, url=https://www.redhat.com, version=17.1.12, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team) Nov 27 03:33:28 localhost podman[86257]: 2025-11-27 08:33:28.102905938 +0000 UTC m=+0.188862884 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, version=17.1.12, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, architecture=x86_64) Nov 27 03:33:28 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:33:28 localhost podman[86255]: 2025-11-27 08:33:28.156827335 +0000 UTC m=+0.251235330 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi) Nov 27 03:33:28 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:33:28 localhost podman[86257]: 2025-11-27 08:33:28.212715965 +0000 UTC m=+0.298672921 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.buildah.version=1.41.4, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, build-date=2025-11-19T00:11:48Z, io.openshift.expose-services=, container_name=ceilometer_agent_compute, managed_by=tripleo_ansible, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:33:28 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:33:28 localhost systemd[1]: tmp-crun.QexxeB.mount: Deactivated successfully. Nov 27 03:33:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:33:30 localhost systemd[1]: tmp-crun.lAmiza.mount: Deactivated successfully. Nov 27 03:33:30 localhost podman[86326]: 2025-11-27 08:33:30.985792621 +0000 UTC m=+0.082208623 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, io.buildah.version=1.41.4, batch=17.1_20251118.1, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:33:31 localhost podman[86326]: 2025-11-27 08:33:31.349982841 +0000 UTC m=+0.446398903 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.openshift.expose-services=, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, version=17.1.12, vcs-type=git, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, distribution-scope=public) Nov 27 03:33:31 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36340 [27/Nov/2025:08:33:30.857] listener listener/metadata 0/0/0/1288/1288 200 130 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-id HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36344 [27/Nov/2025:08:33:32.257] listener listener/metadata 0/0/0/17/17 404 281 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-keys HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36350 [27/Nov/2025:08:33:32.326] listener listener/metadata 0/0/0/12/12 200 130 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-id HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36362 [27/Nov/2025:08:33:32.385] listener listener/metadata 0/0/0/12/12 200 120 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/ami-launch-index HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36368 [27/Nov/2025:08:33:32.434] listener listener/metadata 0/0/0/12/12 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-type HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36376 [27/Nov/2025:08:33:32.490] listener listener/metadata 0/0/0/11/11 200 133 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/local-ipv4 HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36388 [27/Nov/2025:08:33:32.547] listener listener/metadata 0/0/0/13/13 200 134 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-ipv4 HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36404 [27/Nov/2025:08:33:32.598] listener listener/metadata 0/0/0/13/13 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/hostname HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36420 [27/Nov/2025:08:33:32.649] listener listener/metadata 0/0/0/11/11 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/local-hostname HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36430 [27/Nov/2025:08:33:32.699] listener listener/metadata 0/0/0/17/17 404 281 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/user-data HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36446 [27/Nov/2025:08:33:32.753] listener listener/metadata 0/0/0/13/13 200 139 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36460 [27/Nov/2025:08:33:32.793] listener listener/metadata 0/0/0/19/19 200 122 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/ami HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36468 [27/Nov/2025:08:33:32.839] listener listener/metadata 0/0/0/15/15 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/ephemeral0 HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36474 [27/Nov/2025:08:33:32.880] listener listener/metadata 0/0/0/12/12 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/root HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36480 [27/Nov/2025:08:33:32.930] listener listener/metadata 0/0/0/10/10 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-hostname HTTP/1.1" Nov 27 03:33:32 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[86215]: 192.168.0.212:36486 [27/Nov/2025:08:33:32.977] listener listener/metadata 0/0/0/8/8 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.1" Nov 27 03:33:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:33:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:33:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:33:33 localhost podman[86350]: 2025-11-27 08:33:33.986474416 +0000 UTC m=+0.081231416 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:33:34 localhost systemd[1]: tmp-crun.WdylRA.mount: Deactivated successfully. Nov 27 03:33:34 localhost podman[86351]: 2025-11-27 08:33:34.058037169 +0000 UTC m=+0.150282421 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, tcib_managed=true, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, batch=17.1_20251118.1, distribution-scope=public, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:33:34 localhost podman[86350]: 2025-11-27 08:33:34.061842873 +0000 UTC m=+0.156599913 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, io.buildah.version=1.41.4, vcs-type=git, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, release=1761123044) Nov 27 03:33:34 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:33:34 localhost podman[86351]: 2025-11-27 08:33:34.083783486 +0000 UTC m=+0.176028748 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, url=https://www.redhat.com, container_name=ovn_controller, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z) Nov 27 03:33:34 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:33:34 localhost podman[86352]: 2025-11-27 08:33:34.141228407 +0000 UTC m=+0.230919870 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, container_name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.openshift.expose-services=, config_id=tripleo_step5, managed_by=tripleo_ansible, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:33:34 localhost podman[86352]: 2025-11-27 08:33:34.166819569 +0000 UTC m=+0.256511022 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4) Nov 27 03:33:34 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:33:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:33:42 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:33:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:33:42 localhost recover_tripleo_nova_virtqemud[86425]: 63639 Nov 27 03:33:42 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:33:42 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:33:42 localhost systemd[1]: tmp-crun.QcYR1T.mount: Deactivated successfully. Nov 27 03:33:43 localhost podman[86417]: 2025-11-27 08:33:43.001801222 +0000 UTC m=+0.092573152 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, distribution-scope=public, release=1761123044, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, container_name=iscsid, tcib_managed=true) Nov 27 03:33:43 localhost podman[86417]: 2025-11-27 08:33:43.037929418 +0000 UTC m=+0.128701368 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible, vcs-type=git, release=1761123044, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, container_name=iscsid, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:44:13Z, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container) Nov 27 03:33:43 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:33:43 localhost podman[86419]: 2025-11-27 08:33:43.059417979 +0000 UTC m=+0.142508612 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_id=tripleo_step3, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, container_name=collectd) Nov 27 03:33:43 localhost podman[86419]: 2025-11-27 08:33:43.06985207 +0000 UTC m=+0.152942703 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, version=17.1.12, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vcs-type=git, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, distribution-scope=public) Nov 27 03:33:43 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:33:43 localhost systemd[1]: tmp-crun.TyX6nT.mount: Deactivated successfully. Nov 27 03:33:44 localhost snmpd[71019]: empty variable list in _query Nov 27 03:33:44 localhost snmpd[71019]: empty variable list in _query Nov 27 03:33:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:33:53 localhost systemd[1]: tmp-crun.fyTZqs.mount: Deactivated successfully. Nov 27 03:33:54 localhost podman[86459]: 2025-11-27 08:33:54.010033047 +0000 UTC m=+0.094715031 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, vcs-type=git, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_id=tripleo_step1, io.buildah.version=1.41.4) Nov 27 03:33:54 localhost podman[86459]: 2025-11-27 08:33:54.206937296 +0000 UTC m=+0.291619320 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, managed_by=tripleo_ansible, container_name=metrics_qdr, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-type=git, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:33:54 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:33:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:33:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:33:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:33:59 localhost systemd[1]: tmp-crun.8U5I0f.mount: Deactivated successfully. Nov 27 03:33:59 localhost podman[86536]: 2025-11-27 08:33:59.054044432 +0000 UTC m=+0.140826716 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, batch=17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:33:59 localhost podman[86534]: 2025-11-27 08:33:59.016349824 +0000 UTC m=+0.107876387 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, version=17.1.12, container_name=ceilometer_agent_ipmi, url=https://www.redhat.com, io.openshift.expose-services=, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:33:59 localhost podman[86534]: 2025-11-27 08:33:59.101876434 +0000 UTC m=+0.193402967 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, architecture=x86_64, version=17.1.12, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1) Nov 27 03:33:59 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:33:59 localhost podman[86535]: 2025-11-27 08:33:59.113125298 +0000 UTC m=+0.201726572 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, name=rhosp17/openstack-cron, release=1761123044, version=17.1.12, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, container_name=logrotate_crond) Nov 27 03:33:59 localhost podman[86536]: 2025-11-27 08:33:59.137711062 +0000 UTC m=+0.224493316 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, release=1761123044, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, vendor=Red Hat, Inc., managed_by=tripleo_ansible, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4) Nov 27 03:33:59 localhost podman[86535]: 2025-11-27 08:33:59.151997528 +0000 UTC m=+0.240598792 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, io.buildah.version=1.41.4, release=1761123044, version=17.1.12, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, vcs-type=git, url=https://www.redhat.com, name=rhosp17/openstack-cron, config_id=tripleo_step4, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z) Nov 27 03:33:59 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:33:59 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:34:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:34:01 localhost systemd[1]: tmp-crun.qPCcUl.mount: Deactivated successfully. Nov 27 03:34:02 localhost podman[86602]: 2025-11-27 08:34:02.00218725 +0000 UTC m=+0.090579529 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vcs-type=git, tcib_managed=true, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, io.openshift.expose-services=, release=1761123044, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:34:02 localhost podman[86602]: 2025-11-27 08:34:02.400927284 +0000 UTC m=+0.489319543 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, release=1761123044, architecture=x86_64, tcib_managed=true, io.openshift.expose-services=, version=17.1.12, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:34:02 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:34:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:34:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:34:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:34:04 localhost systemd[1]: tmp-crun.tKnVGp.mount: Deactivated successfully. Nov 27 03:34:04 localhost podman[86626]: 2025-11-27 08:34:04.990608324 +0000 UTC m=+0.088345947 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, container_name=ovn_metadata_agent, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:34:05 localhost podman[86627]: 2025-11-27 08:34:05.010204124 +0000 UTC m=+0.101158284 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.buildah.version=1.41.4, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, io.openshift.expose-services=, tcib_managed=true, name=rhosp17/openstack-ovn-controller, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z) Nov 27 03:34:05 localhost podman[86626]: 2025-11-27 08:34:05.036950107 +0000 UTC m=+0.134687730 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, url=https://www.redhat.com, config_id=tripleo_step4, container_name=ovn_metadata_agent, distribution-scope=public, release=1761123044, io.buildah.version=1.41.4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vcs-type=git) Nov 27 03:34:05 localhost podman[86628]: 2025-11-27 08:34:05.047670507 +0000 UTC m=+0.136087719 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, distribution-scope=public, release=1761123044, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-type=git, batch=17.1_20251118.1, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:34:05 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:34:05 localhost podman[86628]: 2025-11-27 08:34:05.100446272 +0000 UTC m=+0.188863484 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., distribution-scope=public, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, io.buildah.version=1.41.4, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:34:05 localhost podman[86627]: 2025-11-27 08:34:05.109819846 +0000 UTC m=+0.200773986 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, container_name=ovn_controller, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, release=1761123044, vcs-type=git, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller) Nov 27 03:34:05 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:34:05 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:34:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:34:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:34:13 localhost systemd[1]: tmp-crun.3Er8fz.mount: Deactivated successfully. Nov 27 03:34:13 localhost podman[86781]: 2025-11-27 08:34:13.998212033 +0000 UTC m=+0.095916493 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, managed_by=tripleo_ansible, tcib_managed=true, release=1761123044, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-collectd-container) Nov 27 03:34:14 localhost podman[86781]: 2025-11-27 08:34:14.009026996 +0000 UTC m=+0.106731446 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vendor=Red Hat, Inc., vcs-type=git, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, container_name=collectd, config_id=tripleo_step3, tcib_managed=true, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:34:14 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:34:14 localhost podman[86782]: 2025-11-27 08:34:14.093318753 +0000 UTC m=+0.185088892 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, url=https://www.redhat.com, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, vcs-type=git, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, tcib_managed=true, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044) Nov 27 03:34:14 localhost podman[86782]: 2025-11-27 08:34:14.104282909 +0000 UTC m=+0.196053008 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, distribution-scope=public, url=https://www.redhat.com, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, tcib_managed=true, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:34:14 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:34:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:34:25 localhost systemd[1]: tmp-crun.ehIqbf.mount: Deactivated successfully. Nov 27 03:34:25 localhost podman[86822]: 2025-11-27 08:34:25.242591199 +0000 UTC m=+0.335730822 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, url=https://www.redhat.com, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:34:25 localhost podman[86822]: 2025-11-27 08:34:25.425822539 +0000 UTC m=+0.518962172 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, version=17.1.12, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-qdrouterd, vcs-type=git, url=https://www.redhat.com, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, architecture=x86_64, config_id=tripleo_step1, distribution-scope=public, maintainer=OpenStack TripleO Team) Nov 27 03:34:25 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:34:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:34:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:34:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:34:29 localhost podman[86851]: 2025-11-27 08:34:29.992457068 +0000 UTC m=+0.083734473 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, architecture=x86_64, release=1761123044, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, version=17.1.12, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, config_id=tripleo_step4, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible) Nov 27 03:34:30 localhost systemd[1]: tmp-crun.MivK74.mount: Deactivated successfully. Nov 27 03:34:30 localhost podman[86851]: 2025-11-27 08:34:30.064131045 +0000 UTC m=+0.155408440 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-type=git, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:34:30 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:34:30 localhost podman[86852]: 2025-11-27 08:34:30.066624283 +0000 UTC m=+0.151520336 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, url=https://www.redhat.com, distribution-scope=public, container_name=logrotate_crond, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, vcs-type=git, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:34:30 localhost podman[86853]: 2025-11-27 08:34:30.157824096 +0000 UTC m=+0.240922530 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, vendor=Red Hat, Inc., config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, url=https://www.redhat.com, distribution-scope=public, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team) Nov 27 03:34:30 localhost podman[86853]: 2025-11-27 08:34:30.187269972 +0000 UTC m=+0.270368386 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, release=1761123044, architecture=x86_64, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:34:30 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:34:30 localhost podman[86852]: 2025-11-27 08:34:30.200716655 +0000 UTC m=+0.285612748 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, vendor=Red Hat, Inc., io.openshift.expose-services=, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:34:30 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:34:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:34:32 localhost podman[86922]: 2025-11-27 08:34:32.985375734 +0000 UTC m=+0.080506535 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, release=1761123044, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, architecture=x86_64, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, tcib_managed=true, vcs-type=git, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:34:33 localhost podman[86922]: 2025-11-27 08:34:33.371509557 +0000 UTC m=+0.466640358 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.buildah.version=1.41.4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target) Nov 27 03:34:33 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:34:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:34:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:34:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:34:35 localhost systemd[1]: tmp-crun.TSwbEH.mount: Deactivated successfully. Nov 27 03:34:36 localhost podman[86947]: 2025-11-27 08:34:36.003455352 +0000 UTC m=+0.085872242 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, release=1761123044, vcs-type=git, managed_by=tripleo_ansible, io.openshift.expose-services=, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:34:36 localhost podman[86947]: 2025-11-27 08:34:36.026833703 +0000 UTC m=+0.109250633 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, vcs-type=git, url=https://www.redhat.com, config_id=tripleo_step4) Nov 27 03:34:36 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:34:36 localhost systemd[1]: tmp-crun.bkZPa7.mount: Deactivated successfully. Nov 27 03:34:36 localhost podman[86946]: 2025-11-27 08:34:36.104261215 +0000 UTC m=+0.192058900 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, distribution-scope=public, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, tcib_managed=true, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:34:36 localhost podman[86948]: 2025-11-27 08:34:36.15217476 +0000 UTC m=+0.231748833 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, architecture=x86_64, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:34:36 localhost podman[86948]: 2025-11-27 08:34:36.179882989 +0000 UTC m=+0.259457052 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, distribution-scope=public, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, config_id=tripleo_step5, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute) Nov 27 03:34:36 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:34:36 localhost podman[86946]: 2025-11-27 08:34:36.205017918 +0000 UTC m=+0.292815643 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vendor=Red Hat, Inc.) Nov 27 03:34:36 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:34:38 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #43. Immutable memtables: 0. Nov 27 03:34:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:34:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:34:44 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:34:44 localhost recover_tripleo_nova_virtqemud[87033]: 63639 Nov 27 03:34:44 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:34:44 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:34:44 localhost systemd[1]: tmp-crun.7uarKT.mount: Deactivated successfully. Nov 27 03:34:45 localhost podman[87021]: 2025-11-27 08:34:45.003666092 +0000 UTC m=+0.098490403 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, vendor=Red Hat, Inc., io.openshift.expose-services=, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible) Nov 27 03:34:45 localhost podman[87021]: 2025-11-27 08:34:45.013277782 +0000 UTC m=+0.108102073 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., architecture=x86_64, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, vcs-type=git, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, tcib_managed=true, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:34:45 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:34:45 localhost podman[87022]: 2025-11-27 08:34:45.098621557 +0000 UTC m=+0.190798066 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, vcs-type=git, release=1761123044, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, managed_by=tripleo_ansible, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, vendor=Red Hat, Inc., architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public) Nov 27 03:34:45 localhost podman[87022]: 2025-11-27 08:34:45.137005005 +0000 UTC m=+0.229181534 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, name=rhosp17/openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., tcib_managed=true) Nov 27 03:34:45 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:34:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:34:55 localhost podman[87063]: 2025-11-27 08:34:55.978313898 +0000 UTC m=+0.076612961 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, batch=17.1_20251118.1, version=17.1.12, distribution-scope=public, release=1761123044, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, io.buildah.version=1.41.4, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd) Nov 27 03:34:56 localhost podman[87063]: 2025-11-27 08:34:56.192851385 +0000 UTC m=+0.291150408 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, distribution-scope=public, version=17.1.12, architecture=x86_64, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com) Nov 27 03:34:56 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:35:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:35:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:35:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:35:01 localhost systemd[1]: tmp-crun.l2M1XK.mount: Deactivated successfully. Nov 27 03:35:01 localhost podman[87139]: 2025-11-27 08:35:01.011278955 +0000 UTC m=+0.097624199 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, release=1761123044, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, vcs-type=git, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vendor=Red Hat, Inc., distribution-scope=public, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, build-date=2025-11-19T00:11:48Z) Nov 27 03:35:01 localhost podman[87137]: 2025-11-27 08:35:01.042705154 +0000 UTC m=+0.132777479 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., release=1761123044, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_id=tripleo_step4, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:35:01 localhost podman[87137]: 2025-11-27 08:35:01.101102002 +0000 UTC m=+0.191174347 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.buildah.version=1.41.4, config_id=tripleo_step4, version=17.1.12, container_name=ceilometer_agent_ipmi, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, vendor=Red Hat, Inc.) Nov 27 03:35:01 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:35:01 localhost podman[87139]: 2025-11-27 08:35:01.121513973 +0000 UTC m=+0.207859247 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, version=17.1.12, batch=17.1_20251118.1, managed_by=tripleo_ansible) Nov 27 03:35:01 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:35:01 localhost podman[87138]: 2025-11-27 08:35:01.103139427 +0000 UTC m=+0.192134413 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, build-date=2025-11-18T22:49:32Z, vcs-type=git, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, container_name=logrotate_crond, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, release=1761123044, batch=17.1_20251118.1) Nov 27 03:35:01 localhost podman[87138]: 2025-11-27 08:35:01.181826803 +0000 UTC m=+0.270821819 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, container_name=logrotate_crond, distribution-scope=public, vcs-type=git, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, url=https://www.redhat.com) Nov 27 03:35:01 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:35:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:35:03 localhost systemd[1]: tmp-crun.IA3Q5O.mount: Deactivated successfully. Nov 27 03:35:03 localhost podman[87211]: 2025-11-27 08:35:03.985377292 +0000 UTC m=+0.081915314 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, architecture=x86_64, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, distribution-scope=public, container_name=nova_migration_target, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, version=17.1.12, release=1761123044, managed_by=tripleo_ansible, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute) Nov 27 03:35:04 localhost podman[87211]: 2025-11-27 08:35:04.352891972 +0000 UTC m=+0.449429954 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, release=1761123044, architecture=x86_64, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, version=17.1.12, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, container_name=nova_migration_target, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:35:04 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:35:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:35:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:35:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:35:06 localhost podman[87234]: 2025-11-27 08:35:06.983707564 +0000 UTC m=+0.080580548 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, vcs-type=git, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, version=17.1.12, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team) Nov 27 03:35:07 localhost systemd[1]: tmp-crun.3kYyXc.mount: Deactivated successfully. Nov 27 03:35:07 localhost podman[87236]: 2025-11-27 08:35:07.043827508 +0000 UTC m=+0.134851074 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, release=1761123044, tcib_managed=true, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team) Nov 27 03:35:07 localhost podman[87234]: 2025-11-27 08:35:07.053927641 +0000 UTC m=+0.150800675 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.expose-services=, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, io.buildah.version=1.41.4, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team) Nov 27 03:35:07 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:35:07 localhost podman[87236]: 2025-11-27 08:35:07.072820802 +0000 UTC m=+0.163844398 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, version=17.1.12, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, managed_by=tripleo_ansible, vcs-type=git, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:35:07 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:35:07 localhost podman[87235]: 2025-11-27 08:35:07.147623193 +0000 UTC m=+0.240367686 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, name=rhosp17/openstack-ovn-controller, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, version=17.1.12, vendor=Red Hat, Inc., io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-type=git, release=1761123044, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, distribution-scope=public, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:35:07 localhost podman[87235]: 2025-11-27 08:35:07.201041886 +0000 UTC m=+0.293786309 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, container_name=ovn_controller, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12) Nov 27 03:35:07 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:35:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:35:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:35:16 localhost systemd[1]: tmp-crun.vaQyVx.mount: Deactivated successfully. Nov 27 03:35:16 localhost podman[87382]: 2025-11-27 08:35:16.011496808 +0000 UTC m=+0.102689645 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=collectd, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:35:16 localhost podman[87383]: 2025-11-27 08:35:16.055688742 +0000 UTC m=+0.145518852 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, io.buildah.version=1.41.4, name=rhosp17/openstack-iscsid, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, container_name=iscsid) Nov 27 03:35:16 localhost podman[87382]: 2025-11-27 08:35:16.056936337 +0000 UTC m=+0.148129164 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, version=17.1.12, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, tcib_managed=true) Nov 27 03:35:16 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:35:16 localhost podman[87383]: 2025-11-27 08:35:16.139184419 +0000 UTC m=+0.229014479 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, distribution-scope=public, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, io.buildah.version=1.41.4, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, container_name=iscsid, com.redhat.component=openstack-iscsid-container, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true) Nov 27 03:35:16 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:35:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:35:26 localhost podman[87421]: 2025-11-27 08:35:26.982414284 +0000 UTC m=+0.079793378 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, version=17.1.12, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd) Nov 27 03:35:27 localhost podman[87421]: 2025-11-27 08:35:27.171003479 +0000 UTC m=+0.268382613 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, architecture=x86_64, release=1761123044, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, vcs-type=git, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., batch=17.1_20251118.1, tcib_managed=true) Nov 27 03:35:27 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:35:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:35:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:35:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:35:31 localhost systemd[1]: tmp-crun.udO0vs.mount: Deactivated successfully. Nov 27 03:35:31 localhost podman[87449]: 2025-11-27 08:35:31.993945811 +0000 UTC m=+0.087910707 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4) Nov 27 03:35:32 localhost podman[87449]: 2025-11-27 08:35:32.032952044 +0000 UTC m=+0.126916910 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, config_id=tripleo_step4, architecture=x86_64, tcib_managed=true, url=https://www.redhat.com, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vendor=Red Hat, Inc., io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-type=git) Nov 27 03:35:32 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:35:32 localhost podman[87450]: 2025-11-27 08:35:32.055779932 +0000 UTC m=+0.145119383 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, version=17.1.12, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z, container_name=ceilometer_agent_compute, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:35:32 localhost podman[87450]: 2025-11-27 08:35:32.086946934 +0000 UTC m=+0.176286405 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, url=https://www.redhat.com, version=17.1.12, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:35:32 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:35:32 localhost podman[87448]: 2025-11-27 08:35:32.144880908 +0000 UTC m=+0.241320111 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, config_id=tripleo_step4, version=17.1.12, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, distribution-scope=public, build-date=2025-11-19T00:12:45Z, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:35:32 localhost podman[87448]: 2025-11-27 08:35:32.208890928 +0000 UTC m=+0.305330141 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, tcib_managed=true, distribution-scope=public) Nov 27 03:35:32 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:35:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:35:34 localhost systemd[1]: tmp-crun.05MdoA.mount: Deactivated successfully. Nov 27 03:35:35 localhost podman[87521]: 2025-11-27 08:35:34.999853348 +0000 UTC m=+0.099813859 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, container_name=nova_migration_target, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team) Nov 27 03:35:35 localhost podman[87521]: 2025-11-27 08:35:35.365466276 +0000 UTC m=+0.465426787 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, io.buildah.version=1.41.4, container_name=nova_migration_target, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, tcib_managed=true, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:35:35 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:35:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:35:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:35:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:35:37 localhost systemd[1]: tmp-crun.oqDv17.mount: Deactivated successfully. Nov 27 03:35:37 localhost podman[87545]: 2025-11-27 08:35:37.980946265 +0000 UTC m=+0.077611348 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, container_name=ovn_metadata_agent, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, managed_by=tripleo_ansible, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 03:35:38 localhost podman[87547]: 2025-11-27 08:35:38.022879588 +0000 UTC m=+0.109487159 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, url=https://www.redhat.com, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, vcs-type=git, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:35:38 localhost podman[87545]: 2025-11-27 08:35:38.04998775 +0000 UTC m=+0.146652813 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, managed_by=tripleo_ansible, url=https://www.redhat.com, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:35:38 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:35:38 localhost podman[87547]: 2025-11-27 08:35:38.104804871 +0000 UTC m=+0.191412452 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, release=1761123044, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5) Nov 27 03:35:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:35:38 localhost podman[87546]: 2025-11-27 08:35:38.191930705 +0000 UTC m=+0.283100120 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, distribution-scope=public, tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4) Nov 27 03:35:38 localhost podman[87546]: 2025-11-27 08:35:38.237074735 +0000 UTC m=+0.328244140 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.buildah.version=1.41.4, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, tcib_managed=true, config_id=tripleo_step4, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:35:38 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:35:41 localhost sshd[87619]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:35:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:35:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:35:46 localhost podman[87622]: 2025-11-27 08:35:46.992670074 +0000 UTC m=+0.080504846 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, config_id=tripleo_step3, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.expose-services=, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, container_name=iscsid, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:35:47 localhost podman[87622]: 2025-11-27 08:35:47.030917117 +0000 UTC m=+0.118751879 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, tcib_managed=true, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, version=17.1.12, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, architecture=x86_64, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible) Nov 27 03:35:47 localhost systemd[1]: tmp-crun.k11L25.mount: Deactivated successfully. Nov 27 03:35:47 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:35:47 localhost podman[87621]: 2025-11-27 08:35:47.05915965 +0000 UTC m=+0.149094429 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, distribution-scope=public, com.redhat.component=openstack-collectd-container, architecture=x86_64, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:35:47 localhost podman[87621]: 2025-11-27 08:35:47.071977947 +0000 UTC m=+0.161912786 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, distribution-scope=public, version=17.1.12, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-collectd-container, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, release=1761123044, tcib_managed=true, name=rhosp17/openstack-collectd, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:35:47 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:35:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:35:57 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:35:57 localhost recover_tripleo_nova_virtqemud[87664]: 63639 Nov 27 03:35:57 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:35:57 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:35:57 localhost systemd[1]: tmp-crun.NDSeK4.mount: Deactivated successfully. Nov 27 03:35:57 localhost podman[87660]: 2025-11-27 08:35:57.996648232 +0000 UTC m=+0.095280315 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vendor=Red Hat, Inc., com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, url=https://www.redhat.com, batch=17.1_20251118.1, config_id=tripleo_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, release=1761123044, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:35:58 localhost podman[87660]: 2025-11-27 08:35:58.212853754 +0000 UTC m=+0.311485817 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-type=git, architecture=x86_64, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, tcib_managed=true, url=https://www.redhat.com, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, version=17.1.12, config_id=tripleo_step1, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z) Nov 27 03:35:58 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:36:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:36:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:36:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:36:02 localhost podman[87737]: 2025-11-27 08:36:02.988760976 +0000 UTC m=+0.081937186 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, container_name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1) Nov 27 03:36:03 localhost systemd[1]: tmp-crun.8zpKoQ.mount: Deactivated successfully. Nov 27 03:36:03 localhost podman[87737]: 2025-11-27 08:36:03.047798621 +0000 UTC m=+0.140974771 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, distribution-scope=public, managed_by=tripleo_ansible, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:36:03 localhost podman[87738]: 2025-11-27 08:36:03.047350539 +0000 UTC m=+0.137864387 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, build-date=2025-11-18T22:49:32Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:36:03 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:36:03 localhost podman[87739]: 2025-11-27 08:36:03.102133579 +0000 UTC m=+0.189288125 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, vcs-type=git, io.buildah.version=1.41.4, url=https://www.redhat.com, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, maintainer=OpenStack TripleO Team, tcib_managed=true, distribution-scope=public, io.openshift.expose-services=, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:36:03 localhost podman[87739]: 2025-11-27 08:36:03.130207807 +0000 UTC m=+0.217362343 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, tcib_managed=true, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:36:03 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:36:03 localhost podman[87738]: 2025-11-27 08:36:03.18101862 +0000 UTC m=+0.271532518 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, config_id=tripleo_step4) Nov 27 03:36:03 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:36:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:36:05 localhost podman[87809]: 2025-11-27 08:36:05.981011244 +0000 UTC m=+0.078694267 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.buildah.version=1.41.4, release=1761123044, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, tcib_managed=true, version=17.1.12, config_id=tripleo_step4) Nov 27 03:36:06 localhost podman[87809]: 2025-11-27 08:36:06.380015134 +0000 UTC m=+0.477698177 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, container_name=nova_migration_target, architecture=x86_64, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, version=17.1.12) Nov 27 03:36:06 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:36:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:36:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:36:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:36:08 localhost systemd[1]: tmp-crun.pfM0kU.mount: Deactivated successfully. Nov 27 03:36:09 localhost systemd[1]: tmp-crun.YSRrYy.mount: Deactivated successfully. Nov 27 03:36:09 localhost podman[87834]: 2025-11-27 08:36:09.013385715 +0000 UTC m=+0.097783223 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, version=17.1.12, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, architecture=x86_64, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step5, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, name=rhosp17/openstack-nova-compute) Nov 27 03:36:09 localhost podman[87832]: 2025-11-27 08:36:08.980197198 +0000 UTC m=+0.071000189 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, vcs-type=git, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, architecture=x86_64, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, tcib_managed=true, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, batch=17.1_20251118.1, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible) Nov 27 03:36:09 localhost podman[87832]: 2025-11-27 08:36:09.064116046 +0000 UTC m=+0.154918997 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.expose-services=, vcs-type=git, version=17.1.12, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, architecture=x86_64, build-date=2025-11-19T00:14:25Z, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team) Nov 27 03:36:09 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:36:09 localhost podman[87834]: 2025-11-27 08:36:09.092243476 +0000 UTC m=+0.176640984 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, release=1761123044, container_name=nova_compute, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true) Nov 27 03:36:09 localhost podman[87833]: 2025-11-27 08:36:09.068663659 +0000 UTC m=+0.152977475 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, tcib_managed=true, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:36:09 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:36:09 localhost podman[87833]: 2025-11-27 08:36:09.148127716 +0000 UTC m=+0.232441542 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, distribution-scope=public, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., batch=17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:36:09 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:36:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:36:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:36:18 localhost podman[87983]: 2025-11-27 08:36:18.004047745 +0000 UTC m=+0.098192624 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, tcib_managed=true, architecture=x86_64, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, container_name=collectd) Nov 27 03:36:18 localhost podman[87984]: 2025-11-27 08:36:18.059464702 +0000 UTC m=+0.150438835 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, build-date=2025-11-18T23:44:13Z, summary=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, distribution-scope=public, vcs-type=git) Nov 27 03:36:18 localhost podman[87984]: 2025-11-27 08:36:18.074363805 +0000 UTC m=+0.165337938 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, distribution-scope=public, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, version=17.1.12, batch=17.1_20251118.1, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-type=git, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:36:18 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:36:18 localhost podman[87983]: 2025-11-27 08:36:18.126068662 +0000 UTC m=+0.220213511 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vcs-type=git, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, url=https://www.redhat.com, architecture=x86_64, container_name=collectd, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, distribution-scope=public, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:36:18 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:36:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:36:28 localhost podman[88022]: 2025-11-27 08:36:28.991594001 +0000 UTC m=+0.089151800 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true) Nov 27 03:36:29 localhost podman[88022]: 2025-11-27 08:36:29.210376092 +0000 UTC m=+0.307933971 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, version=17.1.12, name=rhosp17/openstack-qdrouterd, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, vendor=Red Hat, Inc., managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:36:29 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:36:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:36:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:36:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:36:33 localhost systemd[1]: tmp-crun.cwB19I.mount: Deactivated successfully. Nov 27 03:36:33 localhost podman[88051]: 2025-11-27 08:36:33.997641301 +0000 UTC m=+0.088221105 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, vcs-type=git, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, version=17.1.12, managed_by=tripleo_ansible, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4) Nov 27 03:36:34 localhost podman[88051]: 2025-11-27 08:36:34.025828343 +0000 UTC m=+0.116408157 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., tcib_managed=true, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, version=17.1.12, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.buildah.version=1.41.4) Nov 27 03:36:34 localhost podman[88052]: 2025-11-27 08:36:34.04422519 +0000 UTC m=+0.133388405 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, distribution-scope=public, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, config_id=tripleo_step4, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, managed_by=tripleo_ansible) Nov 27 03:36:34 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:36:34 localhost podman[88052]: 2025-11-27 08:36:34.05274762 +0000 UTC m=+0.141910845 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, name=rhosp17/openstack-cron, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, release=1761123044, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-cron-container, vcs-type=git, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4) Nov 27 03:36:34 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:36:34 localhost podman[88053]: 2025-11-27 08:36:34.10899401 +0000 UTC m=+0.195118683 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, version=17.1.12, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, container_name=ceilometer_agent_compute, vcs-type=git, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, architecture=x86_64) Nov 27 03:36:34 localhost podman[88053]: 2025-11-27 08:36:34.170028109 +0000 UTC m=+0.256152712 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.buildah.version=1.41.4, url=https://www.redhat.com, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, vcs-type=git, release=1761123044, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, tcib_managed=true, io.openshift.expose-services=, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:36:34 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:36:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:36:36 localhost systemd[1]: tmp-crun.YGZZXx.mount: Deactivated successfully. Nov 27 03:36:36 localhost podman[88123]: 2025-11-27 08:36:36.991652147 +0000 UTC m=+0.086071007 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, container_name=nova_migration_target, io.openshift.expose-services=, config_id=tripleo_step4, vcs-type=git, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:36:37 localhost podman[88123]: 2025-11-27 08:36:37.393315509 +0000 UTC m=+0.487734349 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target) Nov 27 03:36:37 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:36:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:36:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:36:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:36:39 localhost systemd[1]: tmp-crun.GN5u2f.mount: Deactivated successfully. Nov 27 03:36:40 localhost podman[88145]: 2025-11-27 08:36:40.004885861 +0000 UTC m=+0.099077057 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, build-date=2025-11-19T00:14:25Z, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, config_id=tripleo_step4, version=17.1.12, url=https://www.redhat.com, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 03:36:40 localhost systemd[1]: tmp-crun.BxrUAT.mount: Deactivated successfully. Nov 27 03:36:40 localhost podman[88146]: 2025-11-27 08:36:40.055640852 +0000 UTC m=+0.146450607 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-18T23:34:05Z, tcib_managed=true, maintainer=OpenStack TripleO Team, container_name=ovn_controller, io.openshift.expose-services=, io.buildah.version=1.41.4, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044) Nov 27 03:36:40 localhost podman[88146]: 2025-11-27 08:36:40.083923847 +0000 UTC m=+0.174733622 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_id=tripleo_step4, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, container_name=ovn_controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, batch=17.1_20251118.1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:36:40 localhost podman[88145]: 2025-11-27 08:36:40.099827737 +0000 UTC m=+0.194018933 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, version=17.1.12, batch=17.1_20251118.1, container_name=ovn_metadata_agent, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, config_id=tripleo_step4, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:36:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:36:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:36:40 localhost podman[88147]: 2025-11-27 08:36:40.107869394 +0000 UTC m=+0.192468972 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., batch=17.1_20251118.1, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, build-date=2025-11-19T00:36:58Z, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 03:36:40 localhost podman[88147]: 2025-11-27 08:36:40.187887625 +0000 UTC m=+0.272487243 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64) Nov 27 03:36:40 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:36:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:36:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:36:49 localhost podman[88217]: 2025-11-27 08:36:49.023190189 +0000 UTC m=+0.121776931 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, architecture=x86_64, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, container_name=collectd, config_id=tripleo_step3, release=1761123044, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, tcib_managed=true, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12) Nov 27 03:36:49 localhost systemd[1]: tmp-crun.Yi5dY9.mount: Deactivated successfully. Nov 27 03:36:49 localhost podman[88218]: 2025-11-27 08:36:49.049586103 +0000 UTC m=+0.142960794 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, container_name=iscsid, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:36:49 localhost podman[88217]: 2025-11-27 08:36:49.062990945 +0000 UTC m=+0.161577657 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, container_name=collectd, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, version=17.1.12, vcs-type=git, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public) Nov 27 03:36:49 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:36:49 localhost podman[88218]: 2025-11-27 08:36:49.087991331 +0000 UTC m=+0.181366012 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, version=17.1.12, distribution-scope=public, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, release=1761123044, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:36:49 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:36:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:36:59 localhost systemd[1]: tmp-crun.WhpQd7.mount: Deactivated successfully. Nov 27 03:36:59 localhost podman[88256]: 2025-11-27 08:36:59.979941333 +0000 UTC m=+0.078981965 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, distribution-scope=public, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, vendor=Red Hat, Inc., url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:37:00 localhost podman[88256]: 2025-11-27 08:37:00.175279451 +0000 UTC m=+0.274320103 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, release=1761123044, vcs-type=git, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, version=17.1.12, tcib_managed=true, container_name=metrics_qdr, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:37:00 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:37:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:37:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:37:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:37:04 localhost systemd[1]: tmp-crun.gZWN1H.mount: Deactivated successfully. Nov 27 03:37:05 localhost podman[88330]: 2025-11-27 08:37:05.005655034 +0000 UTC m=+0.100678571 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, version=17.1.12, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, config_id=tripleo_step4, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:37:05 localhost podman[88332]: 2025-11-27 08:37:05.053585059 +0000 UTC m=+0.143317873 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, maintainer=OpenStack TripleO Team, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., version=17.1.12, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:37:05 localhost podman[88332]: 2025-11-27 08:37:05.085963003 +0000 UTC m=+0.175695897 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible) Nov 27 03:37:05 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:37:05 localhost podman[88331]: 2025-11-27 08:37:05.108924083 +0000 UTC m=+0.200844587 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, name=rhosp17/openstack-cron, architecture=x86_64, container_name=logrotate_crond, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, url=https://www.redhat.com, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:37:05 localhost podman[88330]: 2025-11-27 08:37:05.114090993 +0000 UTC m=+0.209114530 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-type=git, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, vendor=Red Hat, Inc., config_id=tripleo_step4) Nov 27 03:37:05 localhost podman[88331]: 2025-11-27 08:37:05.120060874 +0000 UTC m=+0.211981388 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, tcib_managed=true, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.component=openstack-cron-container) Nov 27 03:37:05 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:37:05 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:37:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:37:07 localhost podman[88404]: 2025-11-27 08:37:07.997982425 +0000 UTC m=+0.095228244 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-type=git) Nov 27 03:37:08 localhost podman[88404]: 2025-11-27 08:37:08.375819364 +0000 UTC m=+0.473065083 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, release=1761123044, vcs-type=git, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, architecture=x86_64, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4) Nov 27 03:37:08 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:37:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:37:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:37:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:37:11 localhost systemd[1]: tmp-crun.w96bJI.mount: Deactivated successfully. Nov 27 03:37:11 localhost podman[88427]: 2025-11-27 08:37:11.008546099 +0000 UTC m=+0.100566677 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:37:11 localhost podman[88428]: 2025-11-27 08:37:11.055480923 +0000 UTC m=+0.145480947 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:37:11 localhost podman[88428]: 2025-11-27 08:37:11.080746273 +0000 UTC m=+0.170746237 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, name=rhosp17/openstack-ovn-controller, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, url=https://www.redhat.com, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:37:11 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:37:11 localhost podman[88427]: 2025-11-27 08:37:11.097784382 +0000 UTC m=+0.189804980 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, architecture=x86_64, tcib_managed=true, maintainer=OpenStack TripleO Team, version=17.1.12, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, container_name=ovn_metadata_agent, io.openshift.expose-services=, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:37:11 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:37:11 localhost podman[88429]: 2025-11-27 08:37:11.150068289 +0000 UTC m=+0.236367383 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, managed_by=tripleo_ansible, io.openshift.expose-services=, release=1761123044, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, distribution-scope=public) Nov 27 03:37:11 localhost podman[88429]: 2025-11-27 08:37:11.205739518 +0000 UTC m=+0.292038562 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, release=1761123044, config_id=tripleo_step5, vcs-type=git, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:37:11 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:37:11 localhost systemd[1]: tmp-crun.7iSBp7.mount: Deactivated successfully. Nov 27 03:37:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:37:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:37:19 localhost podman[88579]: 2025-11-27 08:37:19.995107559 +0000 UTC m=+0.090024514 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, version=17.1.12, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, com.redhat.component=openstack-collectd-container) Nov 27 03:37:20 localhost podman[88579]: 2025-11-27 08:37:20.033095262 +0000 UTC m=+0.128012167 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_id=tripleo_step3, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, container_name=collectd, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1) Nov 27 03:37:20 localhost systemd[1]: tmp-crun.lQUFNM.mount: Deactivated successfully. Nov 27 03:37:20 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:37:20 localhost podman[88580]: 2025-11-27 08:37:20.060180861 +0000 UTC m=+0.154433159 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, build-date=2025-11-18T23:44:13Z, distribution-scope=public, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, io.buildah.version=1.41.4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git) Nov 27 03:37:20 localhost podman[88580]: 2025-11-27 08:37:20.075088552 +0000 UTC m=+0.169340900 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_step3, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, container_name=iscsid, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:37:20 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:37:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:37:30 localhost podman[88617]: 2025-11-27 08:37:30.989494447 +0000 UTC m=+0.083993042 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, container_name=metrics_qdr, io.openshift.expose-services=, architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, version=17.1.12, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com) Nov 27 03:37:31 localhost podman[88617]: 2025-11-27 08:37:31.178837044 +0000 UTC m=+0.273335569 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, version=17.1.12, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, maintainer=OpenStack TripleO Team, config_id=tripleo_step1) Nov 27 03:37:31 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:37:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:37:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:37:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:37:35 localhost podman[88646]: 2025-11-27 08:37:35.992914909 +0000 UTC m=+0.083565881 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, tcib_managed=true, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, distribution-scope=public, io.buildah.version=1.41.4, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:37:36 localhost podman[88646]: 2025-11-27 08:37:36.04792896 +0000 UTC m=+0.138579932 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, managed_by=tripleo_ansible, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.openshift.expose-services=) Nov 27 03:37:36 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:37:36 localhost podman[88647]: 2025-11-27 08:37:36.095872361 +0000 UTC m=+0.185422442 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, architecture=x86_64, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, url=https://www.redhat.com) Nov 27 03:37:36 localhost podman[88647]: 2025-11-27 08:37:36.136946566 +0000 UTC m=+0.226496617 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, batch=17.1_20251118.1, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, name=rhosp17/openstack-cron, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, vendor=Red Hat, Inc., architecture=x86_64, vcs-type=git, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z) Nov 27 03:37:36 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:37:36 localhost podman[88648]: 2025-11-27 08:37:36.047850078 +0000 UTC m=+0.135561180 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team) Nov 27 03:37:36 localhost podman[88648]: 2025-11-27 08:37:36.181948927 +0000 UTC m=+0.269660039 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-type=git, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc.) Nov 27 03:37:36 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:37:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:37:38 localhost podman[88719]: 2025-11-27 08:37:38.996695035 +0000 UTC m=+0.090433025 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_migration_target, version=17.1.12, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, release=1761123044, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, vendor=Red Hat, Inc., batch=17.1_20251118.1) Nov 27 03:37:39 localhost podman[88719]: 2025-11-27 08:37:39.401150002 +0000 UTC m=+0.494887982 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, config_id=tripleo_step4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:37:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:37:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:37:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:37:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:37:42 localhost systemd[1]: tmp-crun.RL0OLI.mount: Deactivated successfully. Nov 27 03:37:42 localhost podman[88742]: 2025-11-27 08:37:42.006635447 +0000 UTC m=+0.096203951 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, release=1761123044, build-date=2025-11-19T00:14:25Z, vcs-type=git, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, io.buildah.version=1.41.4, url=https://www.redhat.com, vendor=Red Hat, Inc.) Nov 27 03:37:42 localhost systemd[1]: tmp-crun.v8pbB1.mount: Deactivated successfully. Nov 27 03:37:42 localhost podman[88743]: 2025-11-27 08:37:42.051478414 +0000 UTC m=+0.140049421 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, io.buildah.version=1.41.4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public) Nov 27 03:37:42 localhost podman[88742]: 2025-11-27 08:37:42.084202334 +0000 UTC m=+0.173770888 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, io.openshift.expose-services=, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, tcib_managed=true) Nov 27 03:37:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:37:42 localhost podman[88744]: 2025-11-27 08:37:42.098423367 +0000 UTC m=+0.183852429 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, architecture=x86_64, managed_by=tripleo_ansible, io.openshift.expose-services=, vendor=Red Hat, Inc., distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, release=1761123044, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, tcib_managed=true) Nov 27 03:37:42 localhost podman[88744]: 2025-11-27 08:37:42.126353649 +0000 UTC m=+0.211782701 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, architecture=x86_64, version=17.1.12, io.openshift.expose-services=, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, release=1761123044) Nov 27 03:37:42 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:37:42 localhost podman[88743]: 2025-11-27 08:37:42.153184452 +0000 UTC m=+0.241755409 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, version=17.1.12, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, distribution-scope=public, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., release=1761123044, container_name=ovn_controller) Nov 27 03:37:42 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:37:44 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:37:44 localhost recover_tripleo_nova_virtqemud[88817]: 63639 Nov 27 03:37:44 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:37:44 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:37:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:37:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:37:50 localhost systemd[1]: tmp-crun.BLyG5q.mount: Deactivated successfully. Nov 27 03:37:51 localhost podman[88819]: 2025-11-27 08:37:50.998723518 +0000 UTC m=+0.090760375 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=iscsid, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, architecture=x86_64, version=17.1.12, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, url=https://www.redhat.com, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid) Nov 27 03:37:51 localhost podman[88819]: 2025-11-27 08:37:51.015081497 +0000 UTC m=+0.107118324 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, distribution-scope=public, container_name=iscsid, release=1761123044, vcs-type=git, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, config_id=tripleo_step3, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12) Nov 27 03:37:51 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:37:51 localhost podman[88818]: 2025-11-27 08:37:50.966169211 +0000 UTC m=+0.065094963 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, batch=17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, distribution-scope=public, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, release=1761123044, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:37:51 localhost podman[88818]: 2025-11-27 08:37:51.096358885 +0000 UTC m=+0.195284607 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, version=17.1.12, distribution-scope=public, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, architecture=x86_64, config_id=tripleo_step3, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, vcs-type=git) Nov 27 03:37:51 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:37:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:37:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 3000.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 413 writes, 1889 keys, 413 commit groups, 1.0 writes per commit group, ingest: 2.85 MB, 0.00 MB/s#012Interval WAL: 413 writes, 120 syncs, 3.44 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:37:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:37:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 3000.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 666 writes, 3016 keys, 666 commit groups, 1.0 writes per commit group, ingest: 4.91 MB, 0.01 MB/s#012Interval WAL: 666 writes, 198 syncs, 3.36 writes per sync, written: 0.00 GB, 0.01 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:38:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:38:02 localhost systemd[1]: tmp-crun.bM2RwU.mount: Deactivated successfully. Nov 27 03:38:02 localhost podman[88878]: 2025-11-27 08:38:02.030777832 +0000 UTC m=+0.117409202 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, batch=17.1_20251118.1, architecture=x86_64, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:38:02 localhost podman[88878]: 2025-11-27 08:38:02.20755445 +0000 UTC m=+0.294185870 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, config_id=tripleo_step1, container_name=metrics_qdr, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, tcib_managed=true, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:38:02 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:38:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:38:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:38:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:38:06 localhost podman[88933]: 2025-11-27 08:38:06.991669419 +0000 UTC m=+0.080719044 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.4, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, container_name=ceilometer_agent_compute, url=https://www.redhat.com, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, architecture=x86_64, version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:38:07 localhost podman[88933]: 2025-11-27 08:38:07.046987247 +0000 UTC m=+0.136036892 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, container_name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, architecture=x86_64) Nov 27 03:38:07 localhost systemd[1]: tmp-crun.hGAfbK.mount: Deactivated successfully. Nov 27 03:38:07 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:38:07 localhost podman[88932]: 2025-11-27 08:38:07.099206383 +0000 UTC m=+0.190585841 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, name=rhosp17/openstack-cron, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:32Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, container_name=logrotate_crond, managed_by=tripleo_ansible, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-type=git, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:38:07 localhost podman[88931]: 2025-11-27 08:38:07.067036227 +0000 UTC m=+0.160715877 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi, release=1761123044, vcs-type=git, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, version=17.1.12, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:38:07 localhost podman[88932]: 2025-11-27 08:38:07.13288737 +0000 UTC m=+0.224266828 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, version=17.1.12, architecture=x86_64, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, vcs-type=git, distribution-scope=public, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:38:07 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:38:07 localhost podman[88931]: 2025-11-27 08:38:07.148059948 +0000 UTC m=+0.241739588 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, io.buildah.version=1.41.4, url=https://www.redhat.com, config_id=tripleo_step4, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, tcib_managed=true, architecture=x86_64, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12) Nov 27 03:38:07 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:38:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:38:09 localhost systemd[85983]: Created slice User Background Tasks Slice. Nov 27 03:38:10 localhost systemd[85983]: Starting Cleanup of User's Temporary Files and Directories... Nov 27 03:38:10 localhost systemd[1]: tmp-crun.Won5Ky.mount: Deactivated successfully. Nov 27 03:38:10 localhost podman[89005]: 2025-11-27 08:38:10.006443181 +0000 UTC m=+0.100214209 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, version=17.1.12, vcs-type=git, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, tcib_managed=true, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:38:10 localhost systemd[85983]: Finished Cleanup of User's Temporary Files and Directories. Nov 27 03:38:10 localhost podman[89005]: 2025-11-27 08:38:10.380953761 +0000 UTC m=+0.474724789 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, managed_by=tripleo_ansible, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, container_name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044) Nov 27 03:38:10 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:38:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:38:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:38:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:38:12 localhost podman[89029]: 2025-11-27 08:38:12.993584938 +0000 UTC m=+0.086936431 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, io.openshift.expose-services=, config_id=tripleo_step4, container_name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:38:13 localhost systemd[1]: tmp-crun.VPMhG4.mount: Deactivated successfully. Nov 27 03:38:13 localhost podman[89031]: 2025-11-27 08:38:13.050784467 +0000 UTC m=+0.138615111 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, version=17.1.12, config_id=tripleo_step5, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, container_name=nova_compute, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:38:13 localhost podman[89029]: 2025-11-27 08:38:13.081386211 +0000 UTC m=+0.174737694 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, version=17.1.12, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, release=1761123044, architecture=x86_64, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, container_name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z) Nov 27 03:38:13 localhost podman[89031]: 2025-11-27 08:38:13.091809492 +0000 UTC m=+0.179640156 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, build-date=2025-11-19T00:36:58Z, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container, release=1761123044, name=rhosp17/openstack-nova-compute, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_id=tripleo_step5, io.openshift.expose-services=, managed_by=tripleo_ansible, url=https://www.redhat.com, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, io.buildah.version=1.41.4) Nov 27 03:38:13 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:38:13 localhost podman[89030]: 2025-11-27 08:38:13.100875276 +0000 UTC m=+0.193195232 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, release=1761123044, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, architecture=x86_64, container_name=ovn_controller, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, io.buildah.version=1.41.4, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:38:13 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:38:13 localhost podman[89030]: 2025-11-27 08:38:13.152947498 +0000 UTC m=+0.245267504 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, distribution-scope=public, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, release=1761123044, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, version=17.1.12, managed_by=tripleo_ansible, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.4, url=https://www.redhat.com, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc.) Nov 27 03:38:13 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:38:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:38:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:38:21 localhost podman[89177]: 2025-11-27 08:38:21.995806879 +0000 UTC m=+0.088681288 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, distribution-scope=public, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4) Nov 27 03:38:22 localhost podman[89177]: 2025-11-27 08:38:22.007958727 +0000 UTC m=+0.100833076 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., vcs-type=git, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, container_name=collectd, com.redhat.component=openstack-collectd-container, version=17.1.12, tcib_managed=true) Nov 27 03:38:22 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:38:22 localhost podman[89178]: 2025-11-27 08:38:22.093936851 +0000 UTC m=+0.184985801 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, config_id=tripleo_step3, release=1761123044, com.redhat.component=openstack-iscsid-container, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:38:22 localhost podman[89178]: 2025-11-27 08:38:22.103696223 +0000 UTC m=+0.194745153 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, architecture=x86_64, batch=17.1_20251118.1, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public) Nov 27 03:38:22 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:38:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:38:32 localhost podman[89216]: 2025-11-27 08:38:32.965686658 +0000 UTC m=+0.065181655 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., distribution-scope=public, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, managed_by=tripleo_ansible, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, architecture=x86_64, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:38:33 localhost podman[89216]: 2025-11-27 08:38:33.147626496 +0000 UTC m=+0.247121423 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, config_id=tripleo_step1, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:38:33 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:38:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:38:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:38:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:38:37 localhost podman[89247]: 2025-11-27 08:38:37.970327725 +0000 UTC m=+0.062385060 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, architecture=x86_64, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, vcs-type=git) Nov 27 03:38:38 localhost systemd[1]: tmp-crun.fcCCvT.mount: Deactivated successfully. Nov 27 03:38:38 localhost podman[89247]: 2025-11-27 08:38:38.011172944 +0000 UTC m=+0.103230249 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, container_name=logrotate_crond, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, managed_by=tripleo_ansible, vcs-type=git, url=https://www.redhat.com, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, distribution-scope=public, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_id=tripleo_step4, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:38:38 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:38:38 localhost podman[89248]: 2025-11-27 08:38:38.058934059 +0000 UTC m=+0.144077799 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, architecture=x86_64, build-date=2025-11-19T00:11:48Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:38:38 localhost podman[89246]: 2025-11-27 08:38:38.012181591 +0000 UTC m=+0.103136227 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vendor=Red Hat, Inc., release=1761123044, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, managed_by=tripleo_ansible, batch=17.1_20251118.1) Nov 27 03:38:38 localhost podman[89246]: 2025-11-27 08:38:38.09500143 +0000 UTC m=+0.185956116 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, managed_by=tripleo_ansible, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, version=17.1.12, io.openshift.expose-services=, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:38:38 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:38:38 localhost podman[89248]: 2025-11-27 08:38:38.118945655 +0000 UTC m=+0.204089395 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, container_name=ceilometer_agent_compute, architecture=x86_64, config_id=tripleo_step4, release=1761123044, build-date=2025-11-19T00:11:48Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:38:38 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:38:38 localhost systemd[1]: tmp-crun.1UjY7A.mount: Deactivated successfully. Nov 27 03:38:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:38:40 localhost systemd[1]: tmp-crun.gOqDVC.mount: Deactivated successfully. Nov 27 03:38:40 localhost podman[89318]: 2025-11-27 08:38:40.998781225 +0000 UTC m=+0.096522120 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, version=17.1.12, release=1761123044, vcs-type=git, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:38:41 localhost podman[89318]: 2025-11-27 08:38:41.381126907 +0000 UTC m=+0.478867832 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, version=17.1.12, release=1761123044, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4) Nov 27 03:38:41 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:38:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:38:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:38:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:38:43 localhost podman[89342]: 2025-11-27 08:38:43.986626751 +0000 UTC m=+0.077258860 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, version=17.1.12, batch=17.1_20251118.1, vcs-type=git, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:38:44 localhost systemd[1]: tmp-crun.B12yLK.mount: Deactivated successfully. Nov 27 03:38:44 localhost podman[89342]: 2025-11-27 08:38:44.053050149 +0000 UTC m=+0.143682258 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, tcib_managed=true, url=https://www.redhat.com, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:38:44 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:38:44 localhost podman[89344]: 2025-11-27 08:38:44.101253656 +0000 UTC m=+0.187225979 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, distribution-scope=public, managed_by=tripleo_ansible, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, url=https://www.redhat.com, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5) Nov 27 03:38:44 localhost podman[89343]: 2025-11-27 08:38:44.055050843 +0000 UTC m=+0.142432695 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, architecture=x86_64, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, container_name=ovn_controller, batch=17.1_20251118.1, version=17.1.12) Nov 27 03:38:44 localhost podman[89344]: 2025-11-27 08:38:44.128211282 +0000 UTC m=+0.214183595 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, vcs-type=git, io.openshift.expose-services=, release=1761123044, version=17.1.12, batch=17.1_20251118.1) Nov 27 03:38:44 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:38:44 localhost podman[89343]: 2025-11-27 08:38:44.138982473 +0000 UTC m=+0.226364315 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, vcs-type=git, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:38:44 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:38:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:38:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:38:52 localhost systemd[1]: tmp-crun.lOgso8.mount: Deactivated successfully. Nov 27 03:38:52 localhost podman[89417]: 2025-11-27 08:38:52.962567586 +0000 UTC m=+0.063182673 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, distribution-scope=public, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_id=tripleo_step3, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, maintainer=OpenStack TripleO Team, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4) Nov 27 03:38:52 localhost podman[89417]: 2025-11-27 08:38:52.975801771 +0000 UTC m=+0.076416818 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, config_id=tripleo_step3, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, version=17.1.12, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, vcs-type=git, release=1761123044, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:38:52 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:38:53 localhost podman[89418]: 2025-11-27 08:38:53.068098466 +0000 UTC m=+0.166114962 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, build-date=2025-11-18T23:44:13Z, architecture=x86_64, io.buildah.version=1.41.4, version=17.1.12, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:38:53 localhost podman[89418]: 2025-11-27 08:38:53.105853943 +0000 UTC m=+0.203870419 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, url=https://www.redhat.com, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:38:53 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:39:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:39:03 localhost podman[89500]: 2025-11-27 08:39:03.985719986 +0000 UTC m=+0.079935942 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container, architecture=x86_64, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, batch=17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, config_id=tripleo_step1, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc.) Nov 27 03:39:04 localhost podman[89500]: 2025-11-27 08:39:04.172264208 +0000 UTC m=+0.266480094 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, version=17.1.12, url=https://www.redhat.com, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., config_id=tripleo_step1, batch=17.1_20251118.1, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, vcs-type=git, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:39:04 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:39:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:39:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:39:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:39:08 localhost podman[89527]: 2025-11-27 08:39:08.995956153 +0000 UTC m=+0.089888600 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, tcib_managed=true, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, distribution-scope=public, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, version=17.1.12, container_name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:39:09 localhost podman[89528]: 2025-11-27 08:39:09.050603814 +0000 UTC m=+0.142162597 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, tcib_managed=true, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, container_name=logrotate_crond, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_id=tripleo_step4, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vendor=Red Hat, Inc., io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:39:09 localhost podman[89527]: 2025-11-27 08:39:09.050039389 +0000 UTC m=+0.143971786 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, vcs-type=git, tcib_managed=true, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, distribution-scope=public, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:39:09 localhost podman[89528]: 2025-11-27 08:39:09.083176411 +0000 UTC m=+0.174735194 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, container_name=logrotate_crond, vcs-type=git, url=https://www.redhat.com, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:39:09 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:39:09 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:39:09 localhost podman[89529]: 2025-11-27 08:39:09.166115244 +0000 UTC m=+0.252946921 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., config_id=tripleo_step4, architecture=x86_64, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=) Nov 27 03:39:09 localhost podman[89529]: 2025-11-27 08:39:09.193392217 +0000 UTC m=+0.280223934 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, release=1761123044, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, distribution-scope=public) Nov 27 03:39:09 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:39:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:39:11 localhost podman[89600]: 2025-11-27 08:39:11.989202095 +0000 UTC m=+0.082179682 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, container_name=nova_migration_target, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:39:12 localhost podman[89600]: 2025-11-27 08:39:12.355870905 +0000 UTC m=+0.448848432 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, name=rhosp17/openstack-nova-compute, vcs-type=git, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., url=https://www.redhat.com, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 03:39:12 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:39:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:39:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:39:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:39:14 localhost systemd[1]: tmp-crun.KtV3o4.mount: Deactivated successfully. Nov 27 03:39:14 localhost podman[89623]: 2025-11-27 08:39:14.994826362 +0000 UTC m=+0.087019174 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, tcib_managed=true, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, url=https://www.redhat.com, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, release=1761123044, io.openshift.expose-services=) Nov 27 03:39:15 localhost systemd[1]: tmp-crun.X2VVzv.mount: Deactivated successfully. Nov 27 03:39:15 localhost podman[89625]: 2025-11-27 08:39:15.05156252 +0000 UTC m=+0.138895191 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_id=tripleo_step5, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, container_name=nova_compute, vendor=Red Hat, Inc., batch=17.1_20251118.1) Nov 27 03:39:15 localhost podman[89624]: 2025-11-27 08:39:15.10064721 +0000 UTC m=+0.189785200 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, io.buildah.version=1.41.4, managed_by=tripleo_ansible, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., build-date=2025-11-18T23:34:05Z, version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1) Nov 27 03:39:15 localhost podman[89623]: 2025-11-27 08:39:15.118018238 +0000 UTC m=+0.210211090 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, architecture=x86_64, release=1761123044, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible) Nov 27 03:39:15 localhost podman[89624]: 2025-11-27 08:39:15.126874726 +0000 UTC m=+0.216012776 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_id=tripleo_step4, release=1761123044) Nov 27 03:39:15 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:39:15 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:39:15 localhost podman[89625]: 2025-11-27 08:39:15.153072042 +0000 UTC m=+0.240404703 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, architecture=x86_64, config_id=tripleo_step5, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., url=https://www.redhat.com, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, managed_by=tripleo_ansible, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, release=1761123044, maintainer=OpenStack TripleO Team, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:39:15 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:39:17 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:39:17 localhost recover_tripleo_nova_virtqemud[89742]: 63639 Nov 27 03:39:17 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:39:17 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:39:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:39:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:39:23 localhost podman[89827]: 2025-11-27 08:39:23.99423388 +0000 UTC m=+0.083449648 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, version=17.1.12, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, name=rhosp17/openstack-iscsid, architecture=x86_64, build-date=2025-11-18T23:44:13Z, vcs-type=git, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com) Nov 27 03:39:24 localhost podman[89827]: 2025-11-27 08:39:24.005811711 +0000 UTC m=+0.095027429 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=iscsid, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, vcs-type=git, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, distribution-scope=public, batch=17.1_20251118.1, version=17.1.12, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., tcib_managed=true, build-date=2025-11-18T23:44:13Z, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:39:24 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:39:24 localhost podman[89826]: 2025-11-27 08:39:24.100642444 +0000 UTC m=+0.189143973 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, batch=17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, vcs-type=git, tcib_managed=true, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:39:24 localhost podman[89826]: 2025-11-27 08:39:24.114918608 +0000 UTC m=+0.203420167 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.buildah.version=1.41.4, version=17.1.12, io.openshift.expose-services=, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container) Nov 27 03:39:24 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:39:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:39:34 localhost podman[89866]: 2025-11-27 08:39:34.992462401 +0000 UTC m=+0.090823047 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.openshift.expose-services=, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:39:35 localhost podman[89866]: 2025-11-27 08:39:35.221136205 +0000 UTC m=+0.319496811 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, vcs-type=git, vendor=Red Hat, Inc., release=1761123044, build-date=2025-11-18T22:49:46Z, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, io.buildah.version=1.41.4) Nov 27 03:39:35 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:39:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:39:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:39:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:39:39 localhost systemd[1]: tmp-crun.vmg6qt.mount: Deactivated successfully. Nov 27 03:39:40 localhost podman[89895]: 2025-11-27 08:39:40.000981619 +0000 UTC m=+0.090861347 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.openshift.expose-services=, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, tcib_managed=true, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, architecture=x86_64, distribution-scope=public, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:12:45Z) Nov 27 03:39:40 localhost podman[89895]: 2025-11-27 08:39:40.056707739 +0000 UTC m=+0.146587467 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:39:40 localhost podman[89896]: 2025-11-27 08:39:40.055861876 +0000 UTC m=+0.142494397 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, release=1761123044, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, com.redhat.component=openstack-cron-container, config_id=tripleo_step4, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron) Nov 27 03:39:40 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:39:40 localhost podman[89897]: 2025-11-27 08:39:40.105481052 +0000 UTC m=+0.189564414 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, batch=17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, architecture=x86_64, release=1761123044, config_id=tripleo_step4, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true) Nov 27 03:39:40 localhost podman[89897]: 2025-11-27 08:39:40.135054488 +0000 UTC m=+0.219137850 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, architecture=x86_64, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_id=tripleo_step4, batch=17.1_20251118.1, url=https://www.redhat.com, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:39:40 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:39:40 localhost podman[89896]: 2025-11-27 08:39:40.191265461 +0000 UTC m=+0.277897942 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, container_name=logrotate_crond, batch=17.1_20251118.1, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=) Nov 27 03:39:40 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:39:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:39:42 localhost podman[89966]: 2025-11-27 08:39:42.981017246 +0000 UTC m=+0.079775478 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, architecture=x86_64, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, distribution-scope=public, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:39:43 localhost podman[89966]: 2025-11-27 08:39:43.336193837 +0000 UTC m=+0.434952039 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, url=https://www.redhat.com, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4) Nov 27 03:39:43 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:39:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:39:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:39:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:39:46 localhost systemd[1]: tmp-crun.i0uILN.mount: Deactivated successfully. Nov 27 03:39:46 localhost podman[89989]: 2025-11-27 08:39:46.013910997 +0000 UTC m=+0.107980397 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, tcib_managed=true, io.openshift.expose-services=, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, config_id=tripleo_step4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc.) Nov 27 03:39:46 localhost podman[89989]: 2025-11-27 08:39:46.091513886 +0000 UTC m=+0.185583306 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., url=https://www.redhat.com, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, architecture=x86_64, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:39:46 localhost systemd[1]: tmp-crun.YMkeZn.mount: Deactivated successfully. Nov 27 03:39:46 localhost podman[89991]: 2025-11-27 08:39:46.110300152 +0000 UTC m=+0.195605037 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=nova_compute, maintainer=OpenStack TripleO Team, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, tcib_managed=true, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, batch=17.1_20251118.1, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container) Nov 27 03:39:46 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:39:46 localhost podman[89991]: 2025-11-27 08:39:46.143867676 +0000 UTC m=+0.229172561 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., config_id=tripleo_step5, managed_by=tripleo_ansible, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, url=https://www.redhat.com, container_name=nova_compute, io.buildah.version=1.41.4, architecture=x86_64, vcs-type=git) Nov 27 03:39:46 localhost podman[89990]: 2025-11-27 08:39:46.156423373 +0000 UTC m=+0.246527507 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, maintainer=OpenStack TripleO Team, distribution-scope=public, release=1761123044, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12) Nov 27 03:39:46 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:39:46 localhost podman[89990]: 2025-11-27 08:39:46.212790611 +0000 UTC m=+0.302894755 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, managed_by=tripleo_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, release=1761123044, tcib_managed=true, batch=17.1_20251118.1) Nov 27 03:39:46 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:39:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:39:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:39:54 localhost systemd[1]: tmp-crun.HYHgP4.mount: Deactivated successfully. Nov 27 03:39:55 localhost podman[90065]: 2025-11-27 08:39:55.000119799 +0000 UTC m=+0.093635163 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, tcib_managed=true, io.buildah.version=1.41.4, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, architecture=x86_64, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, build-date=2025-11-18T22:51:28Z, distribution-scope=public, url=https://www.redhat.com, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team) Nov 27 03:39:55 localhost podman[90065]: 2025-11-27 08:39:55.011764412 +0000 UTC m=+0.105279746 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-type=git, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, url=https://www.redhat.com, config_id=tripleo_step3) Nov 27 03:39:55 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:39:55 localhost podman[90066]: 2025-11-27 08:39:55.085222639 +0000 UTC m=+0.176221955 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, container_name=iscsid, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true) Nov 27 03:39:55 localhost podman[90066]: 2025-11-27 08:39:55.102021141 +0000 UTC m=+0.193020467 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, url=https://www.redhat.com, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, architecture=x86_64, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, vcs-type=git) Nov 27 03:39:55 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:40:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:40:05 localhost systemd[1]: tmp-crun.VHF3Hr.mount: Deactivated successfully. Nov 27 03:40:05 localhost podman[90149]: 2025-11-27 08:40:05.998106542 +0000 UTC m=+0.096229762 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, architecture=x86_64, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, version=17.1.12, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.buildah.version=1.41.4, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1) Nov 27 03:40:06 localhost podman[90149]: 2025-11-27 08:40:06.189989427 +0000 UTC m=+0.288112637 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, io.buildah.version=1.41.4, architecture=x86_64, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vendor=Red Hat, Inc., release=1761123044, managed_by=tripleo_ansible, url=https://www.redhat.com, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:40:06 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:40:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:40:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:40:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:40:11 localhost podman[90178]: 2025-11-27 08:40:10.999720027 +0000 UTC m=+0.084366152 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, vcs-type=git, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, tcib_managed=true, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, config_id=tripleo_step4, url=https://www.redhat.com) Nov 27 03:40:11 localhost systemd[1]: tmp-crun.YGtUBG.mount: Deactivated successfully. Nov 27 03:40:11 localhost podman[90180]: 2025-11-27 08:40:11.072509546 +0000 UTC m=+0.152362361 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.openshift.expose-services=, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.buildah.version=1.41.4, version=17.1.12, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, architecture=x86_64) Nov 27 03:40:11 localhost podman[90180]: 2025-11-27 08:40:11.128895234 +0000 UTC m=+0.208747999 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, distribution-scope=public, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, architecture=x86_64, io.buildah.version=1.41.4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:40:11 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:40:11 localhost podman[90179]: 2025-11-27 08:40:11.170029522 +0000 UTC m=+0.252287732 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., io.buildah.version=1.41.4, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.openshift.expose-services=, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron) Nov 27 03:40:11 localhost podman[90178]: 2025-11-27 08:40:11.187845531 +0000 UTC m=+0.272491686 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, vcs-type=git, distribution-scope=public, batch=17.1_20251118.1, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., url=https://www.redhat.com, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:40:11 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:40:11 localhost podman[90179]: 2025-11-27 08:40:11.208095777 +0000 UTC m=+0.290353987 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, version=17.1.12, build-date=2025-11-18T22:49:32Z, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, architecture=x86_64, vcs-type=git, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:40:11 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:40:11 localhost systemd[1]: tmp-crun.agi1Mj.mount: Deactivated successfully. Nov 27 03:40:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:40:13 localhost podman[90253]: 2025-11-27 08:40:13.995715714 +0000 UTC m=+0.081134455 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, container_name=nova_migration_target, version=17.1.12, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, name=rhosp17/openstack-nova-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:40:14 localhost sshd[90274]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:40:14 localhost sshd[90277]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:40:14 localhost podman[90253]: 2025-11-27 08:40:14.371410986 +0000 UTC m=+0.456829777 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, version=17.1.12, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., container_name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, vcs-type=git) Nov 27 03:40:14 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:40:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:40:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:40:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:40:16 localhost podman[90279]: 2025-11-27 08:40:16.999008066 +0000 UTC m=+0.085585735 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, distribution-scope=public, container_name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, io.buildah.version=1.41.4) Nov 27 03:40:17 localhost systemd[1]: tmp-crun.vsDiw4.mount: Deactivated successfully. Nov 27 03:40:17 localhost podman[90280]: 2025-11-27 08:40:17.055820645 +0000 UTC m=+0.139121696 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vcs-type=git, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller) Nov 27 03:40:17 localhost podman[90279]: 2025-11-27 08:40:17.067979632 +0000 UTC m=+0.154557271 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_id=tripleo_step4, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, vcs-type=git, tcib_managed=true, batch=17.1_20251118.1, distribution-scope=public, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}) Nov 27 03:40:17 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:40:17 localhost podman[90280]: 2025-11-27 08:40:17.107075394 +0000 UTC m=+0.190376415 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., io.buildah.version=1.41.4, vcs-type=git, com.redhat.component=openstack-ovn-controller-container, release=1761123044, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=) Nov 27 03:40:17 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:40:17 localhost podman[90281]: 2025-11-27 08:40:17.161052688 +0000 UTC m=+0.242223612 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, vcs-type=git, release=1761123044, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., version=17.1.12, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com) Nov 27 03:40:17 localhost podman[90281]: 2025-11-27 08:40:17.196910512 +0000 UTC m=+0.278081446 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, architecture=x86_64, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., vcs-type=git, url=https://www.redhat.com, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044) Nov 27 03:40:17 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:40:24 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:40:24 localhost recover_tripleo_nova_virtqemud[90431]: 63639 Nov 27 03:40:24 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:40:24 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:40:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:40:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:40:25 localhost systemd[1]: tmp-crun.62pTH3.mount: Deactivated successfully. Nov 27 03:40:26 localhost podman[90433]: 2025-11-27 08:40:25.9980994 +0000 UTC m=+0.091265577 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, container_name=iscsid, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.buildah.version=1.41.4, tcib_managed=true, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, managed_by=tripleo_ansible) Nov 27 03:40:26 localhost podman[90433]: 2025-11-27 08:40:26.034083409 +0000 UTC m=+0.127249526 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, architecture=x86_64, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.buildah.version=1.41.4, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.openshift.expose-services=) Nov 27 03:40:26 localhost systemd[1]: tmp-crun.GwciA7.mount: Deactivated successfully. Nov 27 03:40:26 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:40:26 localhost podman[90432]: 2025-11-27 08:40:26.05680031 +0000 UTC m=+0.153197515 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, managed_by=tripleo_ansible, vcs-type=git, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, container_name=collectd, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, distribution-scope=public, io.openshift.expose-services=) Nov 27 03:40:26 localhost podman[90432]: 2025-11-27 08:40:26.070879649 +0000 UTC m=+0.167276864 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, name=rhosp17/openstack-collectd, config_id=tripleo_step3, container_name=collectd, distribution-scope=public, version=17.1.12, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:40:26 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:40:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:40:36 localhost podman[90469]: 2025-11-27 08:40:36.982135069 +0000 UTC m=+0.081570116 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, version=17.1.12, config_id=tripleo_step1, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:40:37 localhost podman[90469]: 2025-11-27 08:40:37.164275332 +0000 UTC m=+0.263710339 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, release=1761123044, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, config_id=tripleo_step1, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true) Nov 27 03:40:37 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:40:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:40:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:40:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:40:42 localhost systemd[1]: tmp-crun.31PJ5c.mount: Deactivated successfully. Nov 27 03:40:42 localhost podman[90498]: 2025-11-27 08:40:42.051941548 +0000 UTC m=+0.145465277 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., config_id=tripleo_step4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:40:42 localhost podman[90499]: 2025-11-27 08:40:42.010388989 +0000 UTC m=+0.097732692 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, build-date=2025-11-18T22:49:32Z, distribution-scope=public, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.buildah.version=1.41.4, version=17.1.12, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_id=tripleo_step4) Nov 27 03:40:42 localhost podman[90498]: 2025-11-27 08:40:42.086893848 +0000 UTC m=+0.180417567 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, version=17.1.12, io.openshift.expose-services=, architecture=x86_64, url=https://www.redhat.com, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:40:42 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:40:42 localhost podman[90499]: 2025-11-27 08:40:42.140879422 +0000 UTC m=+0.228223085 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, tcib_managed=true, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vcs-type=git) Nov 27 03:40:42 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:40:42 localhost podman[90500]: 2025-11-27 08:40:42.112317183 +0000 UTC m=+0.197142308 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, config_id=tripleo_step4, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true, maintainer=OpenStack TripleO Team, distribution-scope=public, vendor=Red Hat, Inc., vcs-type=git, name=rhosp17/openstack-ceilometer-compute, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:40:42 localhost podman[90500]: 2025-11-27 08:40:42.193074656 +0000 UTC m=+0.277899861 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, architecture=x86_64, version=17.1.12, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:40:42 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:40:42 localhost systemd[1]: tmp-crun.4PKggb.mount: Deactivated successfully. Nov 27 03:40:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:40:44 localhost podman[90570]: 2025-11-27 08:40:44.984037364 +0000 UTC m=+0.081192997 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, distribution-scope=public, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, tcib_managed=true, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, maintainer=OpenStack TripleO Team) Nov 27 03:40:45 localhost podman[90570]: 2025-11-27 08:40:45.357049015 +0000 UTC m=+0.454204588 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, build-date=2025-11-19T00:36:58Z, version=17.1.12, tcib_managed=true, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, io.buildah.version=1.41.4, vcs-type=git, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:40:45 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:40:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:40:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:40:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:40:47 localhost systemd[1]: tmp-crun.yC2Ciq.mount: Deactivated successfully. Nov 27 03:40:48 localhost podman[90593]: 2025-11-27 08:40:48.00562354 +0000 UTC m=+0.094840634 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, batch=17.1_20251118.1, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, version=17.1.12, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vendor=Red Hat, Inc., config_id=tripleo_step4, container_name=ovn_metadata_agent, io.buildah.version=1.41.4) Nov 27 03:40:48 localhost podman[90594]: 2025-11-27 08:40:48.05056142 +0000 UTC m=+0.135716574 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, architecture=x86_64, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, managed_by=tripleo_ansible, url=https://www.redhat.com, container_name=ovn_controller, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:40:48 localhost podman[90593]: 2025-11-27 08:40:48.066864849 +0000 UTC m=+0.156081953 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, architecture=x86_64) Nov 27 03:40:48 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:40:48 localhost podman[90594]: 2025-11-27 08:40:48.080197268 +0000 UTC m=+0.165352492 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, version=17.1.12, managed_by=tripleo_ansible, container_name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, release=1761123044, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, tcib_managed=true, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc.) Nov 27 03:40:48 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:40:48 localhost podman[90595]: 2025-11-27 08:40:48.156003698 +0000 UTC m=+0.236994001 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, distribution-scope=public, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:40:48 localhost podman[90595]: 2025-11-27 08:40:48.185871321 +0000 UTC m=+0.266861664 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, config_id=tripleo_step5, container_name=nova_compute, url=https://www.redhat.com, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, distribution-scope=public, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z) Nov 27 03:40:48 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:40:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:40:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:40:56 localhost systemd[1]: tmp-crun.2k2Ka4.mount: Deactivated successfully. Nov 27 03:40:56 localhost podman[90668]: 2025-11-27 08:40:56.988858553 +0000 UTC m=+0.081086854 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, name=rhosp17/openstack-iscsid, version=17.1.12, distribution-scope=public, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-type=git, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=) Nov 27 03:40:57 localhost podman[90668]: 2025-11-27 08:40:57.026401833 +0000 UTC m=+0.118630084 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, architecture=x86_64, url=https://www.redhat.com, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, name=rhosp17/openstack-iscsid, vcs-type=git, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc.) Nov 27 03:40:57 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:40:57 localhost podman[90667]: 2025-11-27 08:40:57.043949625 +0000 UTC m=+0.137383909 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.openshift.expose-services=, managed_by=tripleo_ansible, release=1761123044, architecture=x86_64, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, url=https://www.redhat.com, distribution-scope=public, batch=17.1_20251118.1, vcs-type=git, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:40:57 localhost podman[90667]: 2025-11-27 08:40:57.057361026 +0000 UTC m=+0.150795360 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_id=tripleo_step3, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, container_name=collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044) Nov 27 03:40:57 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:41:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:41:07 localhost podman[90730]: 2025-11-27 08:41:07.985623592 +0000 UTC m=+0.079771518 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, architecture=x86_64, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr) Nov 27 03:41:08 localhost podman[90730]: 2025-11-27 08:41:08.204173475 +0000 UTC m=+0.298321381 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, release=1761123044, batch=17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:41:08 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:41:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:41:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:41:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:41:13 localhost podman[90759]: 2025-11-27 08:41:12.999184596 +0000 UTC m=+0.093267051 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, container_name=ceilometer_agent_ipmi, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:41:13 localhost podman[90760]: 2025-11-27 08:41:13.051841964 +0000 UTC m=+0.143620588 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, build-date=2025-11-18T22:49:32Z, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, batch=17.1_20251118.1, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team) Nov 27 03:41:13 localhost podman[90759]: 2025-11-27 08:41:13.057719702 +0000 UTC m=+0.151802157 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, distribution-scope=public, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vcs-type=git, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, io.buildah.version=1.41.4, architecture=x86_64, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, batch=17.1_20251118.1) Nov 27 03:41:13 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:41:13 localhost podman[90760]: 2025-11-27 08:41:13.085814438 +0000 UTC m=+0.177593052 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, distribution-scope=public, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.component=openstack-cron-container, architecture=x86_64, version=17.1.12, io.buildah.version=1.41.4, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:41:13 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:41:13 localhost podman[90761]: 2025-11-27 08:41:13.109641919 +0000 UTC m=+0.196557022 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_id=tripleo_step4, version=17.1.12, build-date=2025-11-19T00:11:48Z, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, tcib_managed=true, io.buildah.version=1.41.4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044) Nov 27 03:41:13 localhost podman[90761]: 2025-11-27 08:41:13.141696873 +0000 UTC m=+0.228611996 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, release=1761123044, tcib_managed=true, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, batch=17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:11:48Z) Nov 27 03:41:13 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:41:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:41:15 localhost podman[90833]: 2025-11-27 08:41:15.984411193 +0000 UTC m=+0.080749315 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, managed_by=tripleo_ansible, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vendor=Red Hat, Inc., tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute) Nov 27 03:41:16 localhost podman[90833]: 2025-11-27 08:41:16.317302983 +0000 UTC m=+0.413641105 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, config_id=tripleo_step4, version=17.1.12, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, release=1761123044, vcs-type=git, batch=17.1_20251118.1) Nov 27 03:41:16 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:41:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:41:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:41:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:41:18 localhost systemd[1]: tmp-crun.44xPTF.mount: Deactivated successfully. Nov 27 03:41:18 localhost podman[90856]: 2025-11-27 08:41:18.986724438 +0000 UTC m=+0.083524429 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:41:19 localhost podman[90857]: 2025-11-27 08:41:19.044895344 +0000 UTC m=+0.135711523 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, container_name=ovn_controller, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, batch=17.1_20251118.1, tcib_managed=true, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, config_id=tripleo_step4, version=17.1.12) Nov 27 03:41:19 localhost podman[90857]: 2025-11-27 08:41:19.062252202 +0000 UTC m=+0.153068341 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, name=rhosp17/openstack-ovn-controller, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., url=https://www.redhat.com) Nov 27 03:41:19 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:41:19 localhost podman[90856]: 2025-11-27 08:41:19.071923122 +0000 UTC m=+0.168723093 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, batch=17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, architecture=x86_64, release=1761123044, build-date=2025-11-19T00:14:25Z, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.buildah.version=1.41.4, url=https://www.redhat.com, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:41:19 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:41:19 localhost podman[90862]: 2025-11-27 08:41:19.014033034 +0000 UTC m=+0.095214354 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, distribution-scope=public, vcs-type=git, tcib_managed=true, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, release=1761123044, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:41:19 localhost podman[90862]: 2025-11-27 08:41:19.149081159 +0000 UTC m=+0.230262519 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, build-date=2025-11-19T00:36:58Z, distribution-scope=public, container_name=nova_compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:41:19 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:41:19 localhost systemd[1]: tmp-crun.rBAXaK.mount: Deactivated successfully. Nov 27 03:41:22 localhost systemd[1]: tmp-crun.B8Xpj6.mount: Deactivated successfully. Nov 27 03:41:22 localhost podman[91034]: 2025-11-27 08:41:22.014828439 +0000 UTC m=+0.098973475 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, distribution-scope=public, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, GIT_BRANCH=main, name=rhceph, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vcs-type=git, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 03:41:22 localhost podman[91034]: 2025-11-27 08:41:22.108006067 +0000 UTC m=+0.192151133 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, GIT_CLEAN=True, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, name=rhceph, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, distribution-scope=public, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553) Nov 27 03:41:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:41:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:41:27 localhost systemd[1]: tmp-crun.E02Ygv.mount: Deactivated successfully. Nov 27 03:41:28 localhost podman[91180]: 2025-11-27 08:41:28.007315946 +0000 UTC m=+0.097245748 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-type=git, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, container_name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z) Nov 27 03:41:28 localhost podman[91179]: 2025-11-27 08:41:28.050999362 +0000 UTC m=+0.140432721 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, name=rhosp17/openstack-collectd, container_name=collectd, architecture=x86_64, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., distribution-scope=public) Nov 27 03:41:28 localhost podman[91180]: 2025-11-27 08:41:28.074161955 +0000 UTC m=+0.164091787 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, build-date=2025-11-18T23:44:13Z, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, container_name=iscsid, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, release=1761123044, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, io.buildah.version=1.41.4, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.component=openstack-iscsid-container) Nov 27 03:41:28 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:41:28 localhost podman[91179]: 2025-11-27 08:41:28.089854498 +0000 UTC m=+0.179287887 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, managed_by=tripleo_ansible, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vendor=Red Hat, Inc., architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z) Nov 27 03:41:28 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:41:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:41:38 localhost systemd[1]: tmp-crun.tdQGVe.mount: Deactivated successfully. Nov 27 03:41:38 localhost podman[91215]: 2025-11-27 08:41:38.982409243 +0000 UTC m=+0.080740104 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., vcs-type=git, container_name=metrics_qdr, release=1761123044, build-date=2025-11-18T22:49:46Z, architecture=x86_64, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:41:39 localhost podman[91215]: 2025-11-27 08:41:39.204153412 +0000 UTC m=+0.302484273 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, tcib_managed=true, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, vendor=Red Hat, Inc., architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-type=git) Nov 27 03:41:39 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:41:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:41:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:41:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:41:43 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:41:43 localhost recover_tripleo_nova_virtqemud[91265]: 63639 Nov 27 03:41:43 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:41:43 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:41:43 localhost podman[91246]: 2025-11-27 08:41:43.995393603 +0000 UTC m=+0.089770178 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step4) Nov 27 03:41:44 localhost podman[91247]: 2025-11-27 08:41:44.046394436 +0000 UTC m=+0.138732776 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, distribution-scope=public, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, version=17.1.12, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.buildah.version=1.41.4, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:41:44 localhost podman[91246]: 2025-11-27 08:41:44.071542423 +0000 UTC m=+0.165919008 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, io.openshift.expose-services=, release=1761123044, io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-ipmi, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, distribution-scope=public, build-date=2025-11-19T00:12:45Z, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team) Nov 27 03:41:44 localhost podman[91247]: 2025-11-27 08:41:44.081909281 +0000 UTC m=+0.174247581 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, vcs-type=git, release=1761123044, url=https://www.redhat.com, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible) Nov 27 03:41:44 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:41:44 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:41:44 localhost podman[91248]: 2025-11-27 08:41:44.164225768 +0000 UTC m=+0.251879332 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-type=git, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, release=1761123044) Nov 27 03:41:44 localhost podman[91248]: 2025-11-27 08:41:44.191654046 +0000 UTC m=+0.279307590 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., architecture=x86_64, name=rhosp17/openstack-ceilometer-compute, vcs-type=git, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:41:44 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:41:44 localhost systemd[1]: tmp-crun.kYKv0l.mount: Deactivated successfully. Nov 27 03:41:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:41:47 localhost systemd[1]: tmp-crun.vf5FBH.mount: Deactivated successfully. Nov 27 03:41:47 localhost podman[91318]: 2025-11-27 08:41:47.025746284 +0000 UTC m=+0.089053259 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, release=1761123044, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, distribution-scope=public, url=https://www.redhat.com, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, tcib_managed=true, name=rhosp17/openstack-nova-compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_id=tripleo_step4, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 03:41:47 localhost podman[91318]: 2025-11-27 08:41:47.363111385 +0000 UTC m=+0.426418290 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, container_name=nova_migration_target) Nov 27 03:41:47 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:41:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:41:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:41:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:41:49 localhost podman[91341]: 2025-11-27 08:41:49.995190276 +0000 UTC m=+0.085300957 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, config_id=tripleo_step4, container_name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, release=1761123044, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1) Nov 27 03:41:50 localhost podman[91342]: 2025-11-27 08:41:50.046641531 +0000 UTC m=+0.133780442 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., managed_by=tripleo_ansible, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, distribution-scope=public, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:41:50 localhost podman[91341]: 2025-11-27 08:41:50.071914042 +0000 UTC m=+0.162024713 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, release=1761123044, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, architecture=x86_64) Nov 27 03:41:50 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:41:50 localhost podman[91342]: 2025-11-27 08:41:50.098958959 +0000 UTC m=+0.186097870 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, config_id=tripleo_step4, io.buildah.version=1.41.4, io.openshift.expose-services=, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, tcib_managed=true) Nov 27 03:41:50 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:41:50 localhost podman[91343]: 2025-11-27 08:41:50.158756329 +0000 UTC m=+0.243719331 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=nova_compute, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:41:50 localhost podman[91343]: 2025-11-27 08:41:50.195879898 +0000 UTC m=+0.280842900 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, architecture=x86_64, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, tcib_managed=true, version=17.1.12, vcs-type=git, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:41:50 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:41:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:41:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:41:58 localhost systemd[1]: tmp-crun.dZtMuB.mount: Deactivated successfully. Nov 27 03:41:58 localhost podman[91415]: 2025-11-27 08:41:58.986785603 +0000 UTC m=+0.087921598 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, architecture=x86_64) Nov 27 03:41:59 localhost podman[91415]: 2025-11-27 08:41:59.001861958 +0000 UTC m=+0.102997933 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, container_name=collectd, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, url=https://www.redhat.com, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.expose-services=, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3) Nov 27 03:41:59 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:41:59 localhost podman[91416]: 2025-11-27 08:41:59.096317191 +0000 UTC m=+0.190917230 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, tcib_managed=true, io.openshift.expose-services=, architecture=x86_64, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, distribution-scope=public, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible) Nov 27 03:41:59 localhost podman[91416]: 2025-11-27 08:41:59.139816932 +0000 UTC m=+0.234416891 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:44:13Z, tcib_managed=true, com.redhat.component=openstack-iscsid-container, container_name=iscsid, release=1761123044, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:41:59 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:42:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:42:09 localhost systemd[1]: tmp-crun.NNby5F.mount: Deactivated successfully. Nov 27 03:42:10 localhost podman[91476]: 2025-11-27 08:42:10.008739662 +0000 UTC m=+0.098138743 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, architecture=x86_64, container_name=metrics_qdr, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, url=https://www.redhat.com) Nov 27 03:42:10 localhost podman[91476]: 2025-11-27 08:42:10.251959489 +0000 UTC m=+0.341358580 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12, config_id=tripleo_step1, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc.) Nov 27 03:42:10 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:42:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:42:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:42:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:42:15 localhost podman[91508]: 2025-11-27 08:42:15.008600197 +0000 UTC m=+0.088635556 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, tcib_managed=true, vendor=Red Hat, Inc., container_name=ceilometer_agent_compute, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, version=17.1.12, name=rhosp17/openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 03:42:15 localhost systemd[1]: tmp-crun.fjIaVa.mount: Deactivated successfully. Nov 27 03:42:15 localhost podman[91508]: 2025-11-27 08:42:15.050055803 +0000 UTC m=+0.130091162 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:11:48Z, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:42:15 localhost podman[91507]: 2025-11-27 08:42:15.05961693 +0000 UTC m=+0.149295950 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, distribution-scope=public, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, config_id=tripleo_step4, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:42:15 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:42:15 localhost podman[91506]: 2025-11-27 08:42:15.101787325 +0000 UTC m=+0.192394999 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, config_id=tripleo_step4, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, tcib_managed=true, distribution-scope=public, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:42:15 localhost podman[91506]: 2025-11-27 08:42:15.13313946 +0000 UTC m=+0.223747134 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, vcs-type=git, release=1761123044, distribution-scope=public, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, io.openshift.expose-services=, tcib_managed=true, url=https://www.redhat.com, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:42:15 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:42:15 localhost podman[91507]: 2025-11-27 08:42:15.145609035 +0000 UTC m=+0.235288075 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, version=17.1.12, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, managed_by=tripleo_ansible, distribution-scope=public, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-cron, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:42:15 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:42:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:42:17 localhost systemd[1]: tmp-crun.bwiYCL.mount: Deactivated successfully. Nov 27 03:42:18 localhost podman[91578]: 2025-11-27 08:42:18.001281914 +0000 UTC m=+0.092091689 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, build-date=2025-11-19T00:36:58Z, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, managed_by=tripleo_ansible, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, batch=17.1_20251118.1, url=https://www.redhat.com, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.expose-services=) Nov 27 03:42:18 localhost podman[91578]: 2025-11-27 08:42:18.365075647 +0000 UTC m=+0.455885412 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, maintainer=OpenStack TripleO Team, vcs-type=git, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, version=17.1.12, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:42:18 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:42:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:42:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:42:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:42:20 localhost systemd[1]: tmp-crun.fupPTc.mount: Deactivated successfully. Nov 27 03:42:21 localhost podman[91599]: 2025-11-27 08:42:21.003945291 +0000 UTC m=+0.097358581 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, version=17.1.12, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:42:21 localhost podman[91600]: 2025-11-27 08:42:21.075711733 +0000 UTC m=+0.167206072 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, distribution-scope=public, io.buildah.version=1.41.4, url=https://www.redhat.com, tcib_managed=true, architecture=x86_64, version=17.1.12, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, release=1761123044, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:42:21 localhost podman[91601]: 2025-11-27 08:42:21.100913902 +0000 UTC m=+0.188716571 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, distribution-scope=public, container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, config_id=tripleo_step5, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:42:21 localhost podman[91599]: 2025-11-27 08:42:21.117295813 +0000 UTC m=+0.210709103 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, container_name=ovn_metadata_agent, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, architecture=x86_64, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, io.buildah.version=1.41.4, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:42:21 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:42:21 localhost podman[91601]: 2025-11-27 08:42:21.135563714 +0000 UTC m=+0.223366393 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, managed_by=tripleo_ansible, config_id=tripleo_step5, release=1761123044, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container) Nov 27 03:42:21 localhost podman[91600]: 2025-11-27 08:42:21.156910419 +0000 UTC m=+0.248404758 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, name=rhosp17/openstack-ovn-controller, vcs-type=git, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, container_name=ovn_controller, version=17.1.12, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ovn-controller-container, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc.) Nov 27 03:42:21 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:42:21 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:42:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:42:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:42:29 localhost systemd[1]: tmp-crun.jQn6ow.mount: Deactivated successfully. Nov 27 03:42:29 localhost podman[91750]: 2025-11-27 08:42:29.9952557 +0000 UTC m=+0.089565742 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, container_name=iscsid, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:42:30 localhost podman[91749]: 2025-11-27 08:42:30.040134868 +0000 UTC m=+0.137060470 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, release=1761123044, name=rhosp17/openstack-collectd, url=https://www.redhat.com, io.openshift.expose-services=, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, container_name=collectd, batch=17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:42:30 localhost podman[91750]: 2025-11-27 08:42:30.058897534 +0000 UTC m=+0.153207566 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., tcib_managed=true, version=17.1.12, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, managed_by=tripleo_ansible) Nov 27 03:42:30 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:42:30 localhost podman[91749]: 2025-11-27 08:42:30.073835185 +0000 UTC m=+0.170760817 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, release=1761123044, com.redhat.component=openstack-collectd-container, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, name=rhosp17/openstack-collectd, batch=17.1_20251118.1, config_id=tripleo_step3, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:42:30 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:42:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:42:41 localhost podman[91789]: 2025-11-27 08:42:40.998140017 +0000 UTC m=+0.085312878 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, io.buildah.version=1.41.4, architecture=x86_64, release=1761123044, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, build-date=2025-11-18T22:49:46Z) Nov 27 03:42:41 localhost podman[91789]: 2025-11-27 08:42:41.216073933 +0000 UTC m=+0.303246764 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc., distribution-scope=public, config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, version=17.1.12) Nov 27 03:42:41 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:42:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:42:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:42:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:42:45 localhost podman[91821]: 2025-11-27 08:42:45.99788601 +0000 UTC m=+0.084161976 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, vendor=Red Hat, Inc., architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.buildah.version=1.41.4, release=1761123044, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:42:46 localhost systemd[1]: tmp-crun.ig6EFI.mount: Deactivated successfully. Nov 27 03:42:46 localhost podman[91819]: 2025-11-27 08:42:46.058494132 +0000 UTC m=+0.149486255 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, release=1761123044, build-date=2025-11-19T00:12:45Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public) Nov 27 03:42:46 localhost podman[91821]: 2025-11-27 08:42:46.05991541 +0000 UTC m=+0.146191366 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, vendor=Red Hat, Inc., tcib_managed=true, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, container_name=ceilometer_agent_compute, io.openshift.expose-services=) Nov 27 03:42:46 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:42:46 localhost podman[91819]: 2025-11-27 08:42:46.138924067 +0000 UTC m=+0.229916160 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, architecture=x86_64, config_id=tripleo_step4, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044) Nov 27 03:42:46 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:42:46 localhost podman[91820]: 2025-11-27 08:42:46.15313319 +0000 UTC m=+0.242652644 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., version=17.1.12, distribution-scope=public, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, config_id=tripleo_step4, url=https://www.redhat.com, release=1761123044, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, container_name=logrotate_crond, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container) Nov 27 03:42:46 localhost podman[91820]: 2025-11-27 08:42:46.18696327 +0000 UTC m=+0.276482684 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, managed_by=tripleo_ansible, batch=17.1_20251118.1, container_name=logrotate_crond, io.buildah.version=1.41.4, distribution-scope=public, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, build-date=2025-11-18T22:49:32Z, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git) Nov 27 03:42:46 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:42:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:42:48 localhost systemd[1]: tmp-crun.wI2kcC.mount: Deactivated successfully. Nov 27 03:42:49 localhost podman[91892]: 2025-11-27 08:42:49.002189901 +0000 UTC m=+0.097487575 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, release=1761123044, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute, version=17.1.12, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public) Nov 27 03:42:49 localhost podman[91892]: 2025-11-27 08:42:49.384236855 +0000 UTC m=+0.479534489 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., distribution-scope=public, container_name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z) Nov 27 03:42:49 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:42:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:42:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:42:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:42:52 localhost podman[91915]: 2025-11-27 08:42:51.997773756 +0000 UTC m=+0.092420008 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, architecture=x86_64, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, managed_by=tripleo_ansible, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, distribution-scope=public, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:42:52 localhost podman[91915]: 2025-11-27 08:42:52.044902935 +0000 UTC m=+0.139549157 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, architecture=x86_64, io.buildah.version=1.41.4, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-19T00:14:25Z, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vcs-type=git, distribution-scope=public, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent) Nov 27 03:42:52 localhost systemd[1]: tmp-crun.Peq1xj.mount: Deactivated successfully. Nov 27 03:42:52 localhost podman[91916]: 2025-11-27 08:42:52.054965476 +0000 UTC m=+0.146729911 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, release=1761123044, container_name=ovn_controller, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, architecture=x86_64, url=https://www.redhat.com, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, distribution-scope=public) Nov 27 03:42:52 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:42:52 localhost podman[91917]: 2025-11-27 08:42:52.108153837 +0000 UTC m=+0.197154897 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, release=1761123044, vcs-type=git, architecture=x86_64, batch=17.1_20251118.1, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step5, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=nova_compute, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:42:52 localhost podman[91916]: 2025-11-27 08:42:52.112188136 +0000 UTC m=+0.203952541 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, version=17.1.12, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, distribution-scope=public, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:42:52 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:42:52 localhost podman[91917]: 2025-11-27 08:42:52.168973285 +0000 UTC m=+0.257974305 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, tcib_managed=true) Nov 27 03:42:52 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:43:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:43:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:43:00 localhost podman[91990]: 2025-11-27 08:43:00.993809241 +0000 UTC m=+0.081591747 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, release=1761123044, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, vcs-type=git, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, url=https://www.redhat.com, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, version=17.1.12, tcib_managed=true, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}) Nov 27 03:43:01 localhost podman[91990]: 2025-11-27 08:43:01.006648508 +0000 UTC m=+0.094431044 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, release=1761123044, com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}) Nov 27 03:43:01 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:43:01 localhost podman[91991]: 2025-11-27 08:43:01.093765322 +0000 UTC m=+0.179266936 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, version=17.1.12, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, container_name=iscsid, url=https://www.redhat.com, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.openshift.expose-services=, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible) Nov 27 03:43:01 localhost podman[91991]: 2025-11-27 08:43:01.131619472 +0000 UTC m=+0.217121066 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, container_name=iscsid, vendor=Red Hat, Inc., batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, name=rhosp17/openstack-iscsid, com.redhat.component=openstack-iscsid-container, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, tcib_managed=true, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:43:01 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:43:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:43:11 localhost podman[92052]: 2025-11-27 08:43:11.983614435 +0000 UTC m=+0.083752785 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.openshift.expose-services=, container_name=metrics_qdr, version=17.1.12, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, architecture=x86_64, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:43:12 localhost podman[92052]: 2025-11-27 08:43:12.174852914 +0000 UTC m=+0.274991184 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, batch=17.1_20251118.1, container_name=metrics_qdr, managed_by=tripleo_ansible, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, architecture=x86_64, release=1761123044, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd) Nov 27 03:43:12 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:43:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:43:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:43:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:43:16 localhost systemd[1]: tmp-crun.lS3Qyf.mount: Deactivated successfully. Nov 27 03:43:17 localhost podman[92082]: 2025-11-27 08:43:17.000429178 +0000 UTC m=+0.092073049 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, distribution-scope=public, name=rhosp17/openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, release=1761123044, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, io.buildah.version=1.41.4, version=17.1.12, vcs-type=git, com.redhat.component=openstack-cron-container, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:43:17 localhost podman[92082]: 2025-11-27 08:43:17.008800363 +0000 UTC m=+0.100444224 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, tcib_managed=true, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, build-date=2025-11-18T22:49:32Z, architecture=x86_64, url=https://www.redhat.com, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:43:17 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:43:17 localhost podman[92083]: 2025-11-27 08:43:17.05623594 +0000 UTC m=+0.142375513 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=ceilometer_agent_compute, config_id=tripleo_step4, batch=17.1_20251118.1, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, version=17.1.12, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:43:17 localhost podman[92081]: 2025-11-27 08:43:17.093273447 +0000 UTC m=+0.187906189 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, release=1761123044, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, io.buildah.version=1.41.4, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:43:17 localhost podman[92083]: 2025-11-27 08:43:17.115009593 +0000 UTC m=+0.201149186 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, version=17.1.12, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, managed_by=tripleo_ansible, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:43:17 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:43:17 localhost podman[92081]: 2025-11-27 08:43:17.14798228 +0000 UTC m=+0.242614962 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., distribution-scope=public, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:43:17 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:43:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:43:19 localhost systemd[1]: tmp-crun.zwqfLA.mount: Deactivated successfully. Nov 27 03:43:19 localhost podman[92154]: 2025-11-27 08:43:19.961631327 +0000 UTC m=+0.062838782 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, container_name=nova_migration_target, architecture=x86_64, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, version=17.1.12, url=https://www.redhat.com) Nov 27 03:43:20 localhost podman[92154]: 2025-11-27 08:43:20.315118922 +0000 UTC m=+0.416326437 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, version=17.1.12, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.openshift.expose-services=, config_id=tripleo_step4, vcs-type=git, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, release=1761123044, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, distribution-scope=public, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:43:20 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:43:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:43:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:43:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:43:23 localhost podman[92179]: 2025-11-27 08:43:23.198676612 +0000 UTC m=+0.282425623 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, release=1761123044, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, io.buildah.version=1.41.4) Nov 27 03:43:23 localhost podman[92179]: 2025-11-27 08:43:23.215185246 +0000 UTC m=+0.298934257 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, tcib_managed=true, version=17.1.12, config_id=tripleo_step4, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, architecture=x86_64, io.openshift.expose-services=, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:43:23 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:43:23 localhost podman[92178]: 2025-11-27 08:43:23.308086857 +0000 UTC m=+0.397006908 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, build-date=2025-11-19T00:14:25Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=) Nov 27 03:43:23 localhost podman[92180]: 2025-11-27 08:43:23.359237264 +0000 UTC m=+0.440663043 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, distribution-scope=public, tcib_managed=true, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute) Nov 27 03:43:23 localhost podman[92178]: 2025-11-27 08:43:23.37209165 +0000 UTC m=+0.461011771 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, architecture=x86_64, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:43:23 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:43:23 localhost podman[92180]: 2025-11-27 08:43:23.415731764 +0000 UTC m=+0.497157513 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, batch=17.1_20251118.1, config_id=tripleo_step5, vendor=Red Hat, Inc., release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, io.openshift.expose-services=, managed_by=tripleo_ansible, container_name=nova_compute, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:43:23 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:43:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:43:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:43:31 localhost podman[92326]: 2025-11-27 08:43:31.991708964 +0000 UTC m=+0.088115374 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, distribution-scope=public, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-collectd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, container_name=collectd, version=17.1.12, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4) Nov 27 03:43:32 localhost podman[92326]: 2025-11-27 08:43:32.005402241 +0000 UTC m=+0.101808661 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, managed_by=tripleo_ansible, io.openshift.expose-services=, version=17.1.12, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, config_id=tripleo_step3, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, distribution-scope=public, io.buildah.version=1.41.4, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-collectd-container) Nov 27 03:43:32 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:43:32 localhost systemd[1]: tmp-crun.6AvPSU.mount: Deactivated successfully. Nov 27 03:43:32 localhost podman[92327]: 2025-11-27 08:43:32.09712471 +0000 UTC m=+0.192149632 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-type=git, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, io.buildah.version=1.41.4, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, architecture=x86_64, vendor=Red Hat, Inc., release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public) Nov 27 03:43:32 localhost podman[92327]: 2025-11-27 08:43:32.109231197 +0000 UTC m=+0.204256109 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, container_name=iscsid, distribution-scope=public, build-date=2025-11-18T23:44:13Z, vcs-type=git, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, managed_by=tripleo_ansible, url=https://www.redhat.com) Nov 27 03:43:32 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:43:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:43:42 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:43:42 localhost recover_tripleo_nova_virtqemud[92369]: 63639 Nov 27 03:43:42 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:43:42 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:43:42 localhost systemd[1]: tmp-crun.mWTqIB.mount: Deactivated successfully. Nov 27 03:43:43 localhost podman[92366]: 2025-11-27 08:43:43.003041356 +0000 UTC m=+0.095289746 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, version=17.1.12, vcs-type=git, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, config_id=tripleo_step1, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, architecture=x86_64, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:43:43 localhost podman[92366]: 2025-11-27 08:43:43.192951299 +0000 UTC m=+0.285199699 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_step1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, managed_by=tripleo_ansible, url=https://www.redhat.com, vendor=Red Hat, Inc., architecture=x86_64, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12) Nov 27 03:43:43 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:43:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:43:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:43:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:43:47 localhost systemd[1]: tmp-crun.0wMmuT.mount: Deactivated successfully. Nov 27 03:43:47 localhost podman[92399]: 2025-11-27 08:43:47.997208301 +0000 UTC m=+0.087709712 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, managed_by=tripleo_ansible, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-compute, container_name=ceilometer_agent_compute, version=17.1.12, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4) Nov 27 03:43:48 localhost podman[92399]: 2025-11-27 08:43:48.031058132 +0000 UTC m=+0.121559563 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., batch=17.1_20251118.1, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, vcs-type=git, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, distribution-scope=public) Nov 27 03:43:48 localhost systemd[1]: tmp-crun.YY1sau.mount: Deactivated successfully. Nov 27 03:43:48 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:43:48 localhost podman[92398]: 2025-11-27 08:43:48.056753383 +0000 UTC m=+0.147148001 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, io.buildah.version=1.41.4, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, name=rhosp17/openstack-cron) Nov 27 03:43:48 localhost podman[92397]: 2025-11-27 08:43:48.09150899 +0000 UTC m=+0.182084903 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., tcib_managed=true, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, version=17.1.12, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, batch=17.1_20251118.1, architecture=x86_64) Nov 27 03:43:48 localhost podman[92398]: 2025-11-27 08:43:48.11866341 +0000 UTC m=+0.209058038 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vendor=Red Hat, Inc., batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, config_id=tripleo_step4, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=logrotate_crond) Nov 27 03:43:48 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:43:48 localhost podman[92397]: 2025-11-27 08:43:48.170726742 +0000 UTC m=+0.261302665 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, io.buildah.version=1.41.4, version=17.1.12, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, config_id=tripleo_step4, io.openshift.expose-services=, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public) Nov 27 03:43:48 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:43:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:43:50 localhost podman[92467]: 2025-11-27 08:43:50.983032484 +0000 UTC m=+0.081859465 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, batch=17.1_20251118.1, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, config_id=tripleo_step4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:43:51 localhost podman[92467]: 2025-11-27 08:43:51.35105132 +0000 UTC m=+0.449878281 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, vcs-type=git, vendor=Red Hat, Inc., release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, url=https://www.redhat.com, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, version=17.1.12, config_id=tripleo_step4, io.buildah.version=1.41.4, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true) Nov 27 03:43:51 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:43:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:43:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:43:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:43:53 localhost systemd[1]: tmp-crun.VIMEAf.mount: Deactivated successfully. Nov 27 03:43:53 localhost podman[92491]: 2025-11-27 08:43:53.993246822 +0000 UTC m=+0.078147505 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, release=1761123044, build-date=2025-11-18T23:34:05Z, vcs-type=git, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, architecture=x86_64, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:43:54 localhost podman[92492]: 2025-11-27 08:43:54.033209299 +0000 UTC m=+0.111811141 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.openshift.expose-services=, release=1761123044, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, name=rhosp17/openstack-nova-compute, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vendor=Red Hat, Inc., tcib_managed=true, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, batch=17.1_20251118.1, container_name=nova_compute, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:43:54 localhost podman[92491]: 2025-11-27 08:43:54.044968755 +0000 UTC m=+0.129869408 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, distribution-scope=public, name=rhosp17/openstack-ovn-controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, managed_by=tripleo_ansible) Nov 27 03:43:54 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:43:54 localhost podman[92492]: 2025-11-27 08:43:54.092408342 +0000 UTC m=+0.171010194 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_id=tripleo_step5, tcib_managed=true, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 03:43:54 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:43:54 localhost podman[92490]: 2025-11-27 08:43:54.098910037 +0000 UTC m=+0.185299239 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, vcs-type=git, io.openshift.expose-services=, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, container_name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn) Nov 27 03:43:54 localhost podman[92490]: 2025-11-27 08:43:54.183127573 +0000 UTC m=+0.269516815 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, release=1761123044, architecture=x86_64, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12) Nov 27 03:43:54 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:44:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:44:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:44:02 localhost systemd[1]: tmp-crun.8g773S.mount: Deactivated successfully. Nov 27 03:44:02 localhost podman[92560]: 2025-11-27 08:44:02.993715718 +0000 UTC m=+0.088759089 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, managed_by=tripleo_ansible, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, batch=17.1_20251118.1, container_name=collectd, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, version=17.1.12, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:44:03 localhost podman[92560]: 2025-11-27 08:44:03.00233369 +0000 UTC m=+0.097377071 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, container_name=collectd, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, tcib_managed=true, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, version=17.1.12, name=rhosp17/openstack-collectd, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, distribution-scope=public, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vendor=Red Hat, Inc.) Nov 27 03:44:03 localhost podman[92561]: 2025-11-27 08:44:03.036315765 +0000 UTC m=+0.129076785 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, tcib_managed=true, io.buildah.version=1.41.4, architecture=x86_64, build-date=2025-11-18T23:44:13Z, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.openshift.expose-services=) Nov 27 03:44:03 localhost podman[92561]: 2025-11-27 08:44:03.048924105 +0000 UTC m=+0.141685125 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, io.buildah.version=1.41.4, container_name=iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:44:03 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:44:03 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:44:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:44:13 localhost systemd[1]: tmp-crun.a5skNm.mount: Deactivated successfully. Nov 27 03:44:14 localhost podman[92622]: 2025-11-27 08:44:14.004216253 +0000 UTC m=+0.098032400 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.expose-services=, vcs-type=git, managed_by=tripleo_ansible, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, release=1761123044, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, vendor=Red Hat, Inc., tcib_managed=true, url=https://www.redhat.com, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:44:14 localhost podman[92622]: 2025-11-27 08:44:14.231967283 +0000 UTC m=+0.325783390 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, batch=17.1_20251118.1, version=17.1.12, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.expose-services=, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:44:14 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:44:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:44:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:44:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:44:18 localhost podman[92654]: 2025-11-27 08:44:18.985988222 +0000 UTC m=+0.078148734 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, managed_by=tripleo_ansible, version=17.1.12, batch=17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git) Nov 27 03:44:19 localhost systemd[1]: tmp-crun.idcAbB.mount: Deactivated successfully. Nov 27 03:44:19 localhost podman[92655]: 2025-11-27 08:44:19.055913775 +0000 UTC m=+0.144637595 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, version=17.1.12, managed_by=tripleo_ansible, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, build-date=2025-11-19T00:11:48Z, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:44:19 localhost podman[92653]: 2025-11-27 08:44:19.098212463 +0000 UTC m=+0.191377842 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, build-date=2025-11-19T00:12:45Z, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, version=17.1.12, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, url=https://www.redhat.com, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, batch=17.1_20251118.1, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi) Nov 27 03:44:19 localhost podman[92654]: 2025-11-27 08:44:19.118419887 +0000 UTC m=+0.210580439 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:32Z, architecture=x86_64, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, version=17.1.12, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=) Nov 27 03:44:19 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:44:19 localhost podman[92655]: 2025-11-27 08:44:19.141984272 +0000 UTC m=+0.230708152 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, vcs-type=git, vendor=Red Hat, Inc., distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, config_id=tripleo_step4, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12) Nov 27 03:44:19 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:44:19 localhost podman[92653]: 2025-11-27 08:44:19.154747525 +0000 UTC m=+0.247912854 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, vcs-type=git, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:44:19 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:44:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:44:21 localhost systemd[1]: tmp-crun.UtkRFD.mount: Deactivated successfully. Nov 27 03:44:21 localhost podman[92728]: 2025-11-27 08:44:21.993990421 +0000 UTC m=+0.092221103 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.buildah.version=1.41.4, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, architecture=x86_64, version=17.1.12, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_migration_target) Nov 27 03:44:22 localhost podman[92728]: 2025-11-27 08:44:22.412906547 +0000 UTC m=+0.511137169 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, release=1761123044, url=https://www.redhat.com, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.openshift.expose-services=, distribution-scope=public, managed_by=tripleo_ansible, version=17.1.12, config_id=tripleo_step4, tcib_managed=true, vcs-type=git, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team) Nov 27 03:44:22 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:44:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:44:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:44:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:44:24 localhost podman[92751]: 2025-11-27 08:44:24.994694354 +0000 UTC m=+0.089899601 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, container_name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vcs-type=git, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., version=17.1.12, release=1761123044) Nov 27 03:44:25 localhost podman[92751]: 2025-11-27 08:44:25.042656335 +0000 UTC m=+0.137861612 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.buildah.version=1.41.4, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step4, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=) Nov 27 03:44:25 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:44:25 localhost podman[92753]: 2025-11-27 08:44:25.043331323 +0000 UTC m=+0.132561339 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vcs-type=git, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, container_name=nova_compute, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=) Nov 27 03:44:25 localhost podman[92752]: 2025-11-27 08:44:25.101652233 +0000 UTC m=+0.192919884 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, version=17.1.12, io.openshift.expose-services=, tcib_managed=true, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, url=https://www.redhat.com, distribution-scope=public, container_name=ovn_controller, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, architecture=x86_64) Nov 27 03:44:25 localhost podman[92752]: 2025-11-27 08:44:25.126366958 +0000 UTC m=+0.217634599 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-type=git, version=17.1.12, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, release=1761123044, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:44:25 localhost podman[92753]: 2025-11-27 08:44:25.126700307 +0000 UTC m=+0.215930343 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, container_name=nova_compute, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com) Nov 27 03:44:25 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:44:25 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:44:25 localhost systemd[1]: tmp-crun.UTSy1Q.mount: Deactivated successfully. Nov 27 03:44:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:44:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:44:33 localhost systemd[1]: tmp-crun.tZzTFa.mount: Deactivated successfully. Nov 27 03:44:34 localhost podman[92900]: 2025-11-27 08:44:34.005212481 +0000 UTC m=+0.096097058 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, tcib_managed=true, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, container_name=iscsid) Nov 27 03:44:34 localhost podman[92900]: 2025-11-27 08:44:34.039061052 +0000 UTC m=+0.129945629 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, architecture=x86_64, distribution-scope=public, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:44:34 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:44:34 localhost podman[92899]: 2025-11-27 08:44:34.101731149 +0000 UTC m=+0.195375270 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, tcib_managed=true, release=1761123044, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step3, vcs-type=git, distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, batch=17.1_20251118.1, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team) Nov 27 03:44:34 localhost podman[92899]: 2025-11-27 08:44:34.136722311 +0000 UTC m=+0.230366452 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, release=1761123044, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, version=17.1.12, vcs-type=git, batch=17.1_20251118.1, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd) Nov 27 03:44:34 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:44:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:44:45 localhost systemd[1]: tmp-crun.8xNhCO.mount: Deactivated successfully. Nov 27 03:44:45 localhost podman[92939]: 2025-11-27 08:44:45.01232165 +0000 UTC m=+0.108865061 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, distribution-scope=public, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, release=1761123044, version=17.1.12) Nov 27 03:44:45 localhost podman[92939]: 2025-11-27 08:44:45.272179075 +0000 UTC m=+0.368722446 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, distribution-scope=public, config_id=tripleo_step1, container_name=metrics_qdr, architecture=x86_64, batch=17.1_20251118.1) Nov 27 03:44:45 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:44:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:44:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:44:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:44:49 localhost systemd[1]: tmp-crun.GV1hdA.mount: Deactivated successfully. Nov 27 03:44:49 localhost podman[92968]: 2025-11-27 08:44:49.994331357 +0000 UTC m=+0.089487660 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team) Nov 27 03:44:50 localhost podman[92968]: 2025-11-27 08:44:50.027833698 +0000 UTC m=+0.122989991 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, release=1761123044) Nov 27 03:44:50 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:44:50 localhost podman[92970]: 2025-11-27 08:44:50.049781509 +0000 UTC m=+0.137003309 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-19T00:11:48Z, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64) Nov 27 03:44:50 localhost podman[92970]: 2025-11-27 08:44:50.111127211 +0000 UTC m=+0.198349011 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, config_id=tripleo_step4, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true) Nov 27 03:44:50 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:44:50 localhost podman[92969]: 2025-11-27 08:44:50.110594656 +0000 UTC m=+0.198727171 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, release=1761123044, batch=17.1_20251118.1, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, container_name=logrotate_crond, build-date=2025-11-18T22:49:32Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, io.openshift.expose-services=) Nov 27 03:44:50 localhost podman[92969]: 2025-11-27 08:44:50.190903568 +0000 UTC m=+0.279036063 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, vendor=Red Hat, Inc., config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, maintainer=OpenStack TripleO Team, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:44:50 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:44:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:44:52 localhost systemd[1]: tmp-crun.UVLy5c.mount: Deactivated successfully. Nov 27 03:44:52 localhost podman[93040]: 2025-11-27 08:44:52.993107408 +0000 UTC m=+0.091920724 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, config_id=tripleo_step4, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, io.openshift.expose-services=, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team) Nov 27 03:44:53 localhost podman[93040]: 2025-11-27 08:44:53.370961679 +0000 UTC m=+0.469774995 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, build-date=2025-11-19T00:36:58Z, tcib_managed=true, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:44:53 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:44:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:44:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:44:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:44:56 localhost systemd[1]: tmp-crun.ZKUslx.mount: Deactivated successfully. Nov 27 03:44:56 localhost podman[93063]: 2025-11-27 08:44:56.010929532 +0000 UTC m=+0.102281244 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, architecture=x86_64, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, version=17.1.12, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vendor=Red Hat, Inc., release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:44:56 localhost systemd[1]: tmp-crun.tKxknG.mount: Deactivated successfully. Nov 27 03:44:56 localhost podman[93063]: 2025-11-27 08:44:56.06845825 +0000 UTC m=+0.159809972 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, release=1761123044, batch=17.1_20251118.1, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com) Nov 27 03:44:56 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:44:56 localhost podman[93065]: 2025-11-27 08:44:56.073574698 +0000 UTC m=+0.155623710 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vendor=Red Hat, Inc., release=1761123044, io.buildah.version=1.41.4, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, version=17.1.12) Nov 27 03:44:56 localhost podman[93065]: 2025-11-27 08:44:56.15276361 +0000 UTC m=+0.234812652 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, architecture=x86_64, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com) Nov 27 03:44:56 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:44:56 localhost podman[93064]: 2025-11-27 08:44:56.210200176 +0000 UTC m=+0.297115138 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, distribution-scope=public, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, io.openshift.expose-services=, container_name=ovn_controller, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:44:56 localhost podman[93064]: 2025-11-27 08:44:56.258856406 +0000 UTC m=+0.345771358 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, architecture=x86_64, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, vcs-type=git, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, release=1761123044, managed_by=tripleo_ansible, tcib_managed=true, url=https://www.redhat.com) Nov 27 03:44:56 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:45:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:45:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:45:04 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:45:04 localhost recover_tripleo_nova_virtqemud[93148]: 63639 Nov 27 03:45:04 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:45:04 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:45:04 localhost podman[93136]: 2025-11-27 08:45:04.993996247 +0000 UTC m=+0.084430094 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, tcib_managed=true, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, container_name=collectd, build-date=2025-11-18T22:51:28Z, distribution-scope=public, version=17.1.12, config_id=tripleo_step3, architecture=x86_64, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, name=rhosp17/openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:45:05 localhost podman[93136]: 2025-11-27 08:45:05.0059598 +0000 UTC m=+0.096393677 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, container_name=collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, distribution-scope=public, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, architecture=x86_64, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, build-date=2025-11-18T22:51:28Z, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-collectd-container) Nov 27 03:45:05 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:45:05 localhost systemd[1]: tmp-crun.8ZYETd.mount: Deactivated successfully. Nov 27 03:45:05 localhost podman[93137]: 2025-11-27 08:45:05.108811307 +0000 UTC m=+0.195591846 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.openshift.expose-services=, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, vendor=Red Hat, Inc., batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:45:05 localhost podman[93137]: 2025-11-27 08:45:05.121954421 +0000 UTC m=+0.208735020 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, architecture=x86_64, build-date=2025-11-18T23:44:13Z, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, container_name=iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-type=git, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com) Nov 27 03:45:05 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:45:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:45:16 localhost podman[93178]: 2025-11-27 08:45:16.010043728 +0000 UTC m=+0.095467191 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, vcs-type=git, io.openshift.expose-services=, release=1761123044, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, distribution-scope=public, version=17.1.12, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:45:16 localhost podman[93178]: 2025-11-27 08:45:16.212752684 +0000 UTC m=+0.298176077 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, release=1761123044, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, maintainer=OpenStack TripleO Team, vcs-type=git, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:45:16 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:45:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:45:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:45:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:45:20 localhost systemd[1]: tmp-crun.Nshj52.mount: Deactivated successfully. Nov 27 03:45:20 localhost podman[93208]: 2025-11-27 08:45:20.997678366 +0000 UTC m=+0.089882651 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, release=1761123044, distribution-scope=public, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., vcs-type=git) Nov 27 03:45:21 localhost podman[93208]: 2025-11-27 08:45:21.034002924 +0000 UTC m=+0.126207209 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, vcs-type=git, batch=17.1_20251118.1, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:45:21 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:45:21 localhost podman[93209]: 2025-11-27 08:45:21.054492375 +0000 UTC m=+0.141923701 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, distribution-scope=public, tcib_managed=true, name=rhosp17/openstack-cron, architecture=x86_64, container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044) Nov 27 03:45:21 localhost podman[93209]: 2025-11-27 08:45:21.065889072 +0000 UTC m=+0.153320378 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-type=git, maintainer=OpenStack TripleO Team, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, release=1761123044, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., batch=17.1_20251118.1, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, tcib_managed=true, com.redhat.component=openstack-cron-container, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:45:21 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:45:21 localhost podman[93210]: 2025-11-27 08:45:21.16055426 +0000 UTC m=+0.245715385 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, name=rhosp17/openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, io.openshift.expose-services=, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, build-date=2025-11-19T00:11:48Z, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, release=1761123044, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team) Nov 27 03:45:21 localhost podman[93210]: 2025-11-27 08:45:21.192057038 +0000 UTC m=+0.277218173 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, vcs-type=git, managed_by=tripleo_ansible, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, tcib_managed=true, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:45:21 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:45:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:45:23 localhost systemd[1]: tmp-crun.j56msU.mount: Deactivated successfully. Nov 27 03:45:23 localhost podman[93282]: 2025-11-27 08:45:23.975052852 +0000 UTC m=+0.073339775 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, container_name=nova_migration_target, architecture=x86_64, distribution-scope=public, release=1761123044, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4) Nov 27 03:45:24 localhost podman[93282]: 2025-11-27 08:45:24.372035278 +0000 UTC m=+0.470322221 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, distribution-scope=public, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:45:24 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:45:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:45:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:45:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:45:27 localhost podman[93307]: 2025-11-27 08:45:27.004657422 +0000 UTC m=+0.089287095 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.buildah.version=1.41.4, config_id=tripleo_step5, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:45:27 localhost podman[93307]: 2025-11-27 08:45:27.033979641 +0000 UTC m=+0.118609324 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, version=17.1.12, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_id=tripleo_step5, vendor=Red Hat, Inc.) Nov 27 03:45:27 localhost systemd[1]: tmp-crun.Sa55P8.mount: Deactivated successfully. Nov 27 03:45:27 localhost podman[93306]: 2025-11-27 08:45:27.054893574 +0000 UTC m=+0.141073768 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, container_name=ovn_controller, version=17.1.12, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, release=1761123044, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc.) Nov 27 03:45:27 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:45:27 localhost podman[93305]: 2025-11-27 08:45:27.105681071 +0000 UTC m=+0.196779078 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, tcib_managed=true, distribution-scope=public, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, container_name=ovn_metadata_agent, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:45:27 localhost podman[93306]: 2025-11-27 08:45:27.130785047 +0000 UTC m=+0.216965291 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, url=https://www.redhat.com, distribution-scope=public, architecture=x86_64, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:45:27 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:45:27 localhost podman[93305]: 2025-11-27 08:45:27.163888288 +0000 UTC m=+0.254986335 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, tcib_managed=true, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, config_id=tripleo_step4, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:45:27 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:45:27 localhost systemd[1]: tmp-crun.C7xJzh.mount: Deactivated successfully. Nov 27 03:45:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:45:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:45:36 localhost podman[93456]: 2025-11-27 08:45:35.996632659 +0000 UTC m=+0.086260793 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, config_id=tripleo_step3, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, tcib_managed=true, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044, managed_by=tripleo_ansible, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:45:36 localhost podman[93456]: 2025-11-27 08:45:36.01001654 +0000 UTC m=+0.099644654 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, release=1761123044, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step3, distribution-scope=public, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, vendor=Red Hat, Inc., architecture=x86_64) Nov 27 03:45:36 localhost podman[93457]: 2025-11-27 08:45:36.042973047 +0000 UTC m=+0.132689233 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., tcib_managed=true, container_name=iscsid, distribution-scope=public, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:45:36 localhost podman[93457]: 2025-11-27 08:45:36.055972206 +0000 UTC m=+0.145688452 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, build-date=2025-11-18T23:44:13Z, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:45:36 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:45:36 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:45:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:45:46 localhost systemd[1]: tmp-crun.PIUKwX.mount: Deactivated successfully. Nov 27 03:45:47 localhost podman[93493]: 2025-11-27 08:45:47.006464485 +0000 UTC m=+0.097853226 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, version=17.1.12, config_id=tripleo_step1, build-date=2025-11-18T22:49:46Z, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4) Nov 27 03:45:47 localhost podman[93493]: 2025-11-27 08:45:47.246970148 +0000 UTC m=+0.338358909 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, container_name=metrics_qdr, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:45:47 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:45:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:45:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:45:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:45:51 localhost systemd[1]: tmp-crun.4DZNgf.mount: Deactivated successfully. Nov 27 03:45:52 localhost podman[93520]: 2025-11-27 08:45:52.002167123 +0000 UTC m=+0.097282572 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, tcib_managed=true, container_name=ceilometer_agent_ipmi, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc.) Nov 27 03:45:52 localhost podman[93521]: 2025-11-27 08:45:52.040834549 +0000 UTC m=+0.132775993 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, tcib_managed=true, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64) Nov 27 03:45:52 localhost podman[93520]: 2025-11-27 08:45:52.058988967 +0000 UTC m=+0.154104426 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, io.openshift.expose-services=, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, name=rhosp17/openstack-ceilometer-ipmi, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, batch=17.1_20251118.1) Nov 27 03:45:52 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:45:52 localhost podman[93521]: 2025-11-27 08:45:52.078982403 +0000 UTC m=+0.170923887 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., vcs-type=git, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, io.openshift.expose-services=, tcib_managed=true, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public) Nov 27 03:45:52 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:45:52 localhost podman[93522]: 2025-11-27 08:45:52.151033427 +0000 UTC m=+0.240766661 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, build-date=2025-11-19T00:11:48Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, vcs-type=git, managed_by=tripleo_ansible, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, container_name=ceilometer_agent_compute, io.openshift.expose-services=, io.buildah.version=1.41.4) Nov 27 03:45:52 localhost podman[93522]: 2025-11-27 08:45:52.183892477 +0000 UTC m=+0.273625751 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, url=https://www.redhat.com, container_name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:45:52 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:45:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:45:54 localhost systemd[1]: tmp-crun.vYw9uQ.mount: Deactivated successfully. Nov 27 03:45:54 localhost podman[93593]: 2025-11-27 08:45:54.991762604 +0000 UTC m=+0.090498350 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, io.buildah.version=1.41.4, batch=17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:45:55 localhost podman[93593]: 2025-11-27 08:45:55.402960767 +0000 UTC m=+0.501696563 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, container_name=nova_migration_target, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, url=https://www.redhat.com, distribution-scope=public, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, vcs-type=git, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:45:55 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:45:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:45:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:45:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:45:57 localhost systemd[1]: tmp-crun.kgyjk4.mount: Deactivated successfully. Nov 27 03:45:57 localhost podman[93618]: 2025-11-27 08:45:57.999281146 +0000 UTC m=+0.091643829 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, release=1761123044, vendor=Red Hat, Inc., managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, maintainer=OpenStack TripleO Team) Nov 27 03:45:58 localhost podman[93620]: 2025-11-27 08:45:58.049124973 +0000 UTC m=+0.135467895 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, io.openshift.expose-services=, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, architecture=x86_64, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc.) Nov 27 03:45:58 localhost podman[93618]: 2025-11-27 08:45:58.056888862 +0000 UTC m=+0.149251505 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, io.openshift.expose-services=, io.buildah.version=1.41.4, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, batch=17.1_20251118.1, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, vcs-type=git, release=1761123044, vendor=Red Hat, Inc.) Nov 27 03:45:58 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:45:58 localhost podman[93619]: 2025-11-27 08:45:57.975831287 +0000 UTC m=+0.067662126 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, io.buildah.version=1.41.4, url=https://www.redhat.com, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, maintainer=OpenStack TripleO Team, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, version=17.1.12) Nov 27 03:45:58 localhost podman[93620]: 2025-11-27 08:45:58.096148495 +0000 UTC m=+0.182491377 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, container_name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., url=https://www.redhat.com, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, distribution-scope=public) Nov 27 03:45:58 localhost podman[93619]: 2025-11-27 08:45:58.105864305 +0000 UTC m=+0.197695074 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, distribution-scope=public, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, name=rhosp17/openstack-ovn-controller, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, release=1761123044) Nov 27 03:45:58 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:45:58 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:45:58 localhost systemd[1]: tmp-crun.ayVjph.mount: Deactivated successfully. Nov 27 03:46:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:46:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:46:06 localhost systemd[1]: tmp-crun.TKwlm6.mount: Deactivated successfully. Nov 27 03:46:06 localhost podman[93691]: 2025-11-27 08:46:06.995932928 +0000 UTC m=+0.087554629 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, version=17.1.12, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, batch=17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.component=openstack-iscsid-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.buildah.version=1.41.4, name=rhosp17/openstack-iscsid, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:46:07 localhost podman[93691]: 2025-11-27 08:46:07.030833194 +0000 UTC m=+0.122454905 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, version=17.1.12, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., release=1761123044) Nov 27 03:46:07 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:46:07 localhost podman[93690]: 2025-11-27 08:46:07.088582824 +0000 UTC m=+0.183232557 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., container_name=collectd, release=1761123044, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, build-date=2025-11-18T22:51:28Z, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, architecture=x86_64, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:46:07 localhost podman[93690]: 2025-11-27 08:46:07.099055765 +0000 UTC m=+0.193705508 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, config_id=tripleo_step3, architecture=x86_64, url=https://www.redhat.com, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, release=1761123044, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, tcib_managed=true, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container) Nov 27 03:46:07 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:46:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:46:17 localhost podman[93730]: 2025-11-27 08:46:17.981798084 +0000 UTC m=+0.081466207 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, vcs-type=git, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z) Nov 27 03:46:18 localhost podman[93730]: 2025-11-27 08:46:18.197934803 +0000 UTC m=+0.297602926 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, release=1761123044, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, batch=17.1_20251118.1, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, vendor=Red Hat, Inc., config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, architecture=x86_64, vcs-type=git) Nov 27 03:46:18 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:46:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:46:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:46:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:46:23 localhost systemd[1]: tmp-crun.JMQu36.mount: Deactivated successfully. Nov 27 03:46:23 localhost podman[93759]: 2025-11-27 08:46:23.182772227 +0000 UTC m=+0.076666808 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public) Nov 27 03:46:23 localhost systemd[1]: tmp-crun.oL0JAu.mount: Deactivated successfully. Nov 27 03:46:23 localhost podman[93766]: 2025-11-27 08:46:23.260681347 +0000 UTC m=+0.143294595 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vendor=Red Hat, Inc., config_id=tripleo_step4, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 03:46:23 localhost podman[93759]: 2025-11-27 08:46:23.285979196 +0000 UTC m=+0.179873817 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, managed_by=tripleo_ansible) Nov 27 03:46:23 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:46:23 localhost podman[93766]: 2025-11-27 08:46:23.340421796 +0000 UTC m=+0.223035084 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, architecture=x86_64, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:46:23 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:46:23 localhost podman[93760]: 2025-11-27 08:46:23.362390166 +0000 UTC m=+0.249140856 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, architecture=x86_64, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, com.redhat.component=openstack-cron-container, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc.) Nov 27 03:46:23 localhost podman[93760]: 2025-11-27 08:46:23.371308865 +0000 UTC m=+0.258059555 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, container_name=logrotate_crond, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:46:23 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:46:24 localhost systemd[1]: tmp-crun.ozrkob.mount: Deactivated successfully. Nov 27 03:46:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:46:25 localhost podman[93832]: 2025-11-27 08:46:25.986907413 +0000 UTC m=+0.081869787 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, release=1761123044, tcib_managed=true, url=https://www.redhat.com, config_id=tripleo_step4, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:46:26 localhost podman[93832]: 2025-11-27 08:46:26.383949756 +0000 UTC m=+0.478912080 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute, version=17.1.12, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, vcs-type=git, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:46:26 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:46:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:46:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:46:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:46:28 localhost systemd[1]: tmp-crun.fTAcer.mount: Deactivated successfully. Nov 27 03:46:28 localhost podman[93853]: 2025-11-27 08:46:28.998767302 +0000 UTC m=+0.092926944 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, release=1761123044, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:46:29 localhost podman[93853]: 2025-11-27 08:46:29.044832168 +0000 UTC m=+0.138991890 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-type=git, container_name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, vendor=Red Hat, Inc., config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:46:29 localhost podman[93854]: 2025-11-27 08:46:29.06020046 +0000 UTC m=+0.152592244 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, container_name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, distribution-scope=public, build-date=2025-11-18T23:34:05Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true) Nov 27 03:46:29 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:46:29 localhost podman[93854]: 2025-11-27 08:46:29.087974116 +0000 UTC m=+0.180365870 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, config_id=tripleo_step4, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 03:46:29 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:46:29 localhost podman[93855]: 2025-11-27 08:46:29.101095248 +0000 UTC m=+0.189189047 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step5, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:46:29 localhost podman[93855]: 2025-11-27 08:46:29.127486366 +0000 UTC m=+0.215580105 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, url=https://www.redhat.com, vcs-type=git, build-date=2025-11-19T00:36:58Z, release=1761123044, com.redhat.component=openstack-nova-compute-container, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, config_id=tripleo_step5, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., distribution-scope=public, name=rhosp17/openstack-nova-compute, tcib_managed=true, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:46:29 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:46:29 localhost systemd[1]: tmp-crun.HLdrTG.mount: Deactivated successfully. Nov 27 03:46:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:46:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:46:37 localhost systemd[1]: tmp-crun.lqigHp.mount: Deactivated successfully. Nov 27 03:46:38 localhost podman[94001]: 2025-11-27 08:46:38.000119912 +0000 UTC m=+0.083763089 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, batch=17.1_20251118.1, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, managed_by=tripleo_ansible, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, name=rhosp17/openstack-iscsid, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:46:38 localhost systemd[1]: tmp-crun.MPwwoZ.mount: Deactivated successfully. Nov 27 03:46:38 localhost podman[94001]: 2025-11-27 08:46:38.044051971 +0000 UTC m=+0.127695148 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, release=1761123044, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, com.redhat.component=openstack-iscsid-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible, container_name=iscsid, url=https://www.redhat.com) Nov 27 03:46:38 localhost podman[94000]: 2025-11-27 08:46:38.051772948 +0000 UTC m=+0.139061202 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, version=17.1.12, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, vcs-type=git, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z) Nov 27 03:46:38 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:46:38 localhost podman[94000]: 2025-11-27 08:46:38.088867683 +0000 UTC m=+0.176155927 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.buildah.version=1.41.4, config_id=tripleo_step3, architecture=x86_64, distribution-scope=public, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, build-date=2025-11-18T22:51:28Z, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044) Nov 27 03:46:38 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:46:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:46:48 localhost systemd[1]: tmp-crun.DQABUC.mount: Deactivated successfully. Nov 27 03:46:48 localhost podman[94037]: 2025-11-27 08:46:48.988107472 +0000 UTC m=+0.084638603 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, architecture=x86_64, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, container_name=metrics_qdr, release=1761123044, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git) Nov 27 03:46:49 localhost podman[94037]: 2025-11-27 08:46:49.203354376 +0000 UTC m=+0.299885467 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 03:46:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:46:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:46:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:46:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:46:53 localhost systemd[1]: tmp-crun.erFjNo.mount: Deactivated successfully. Nov 27 03:46:53 localhost podman[94067]: 2025-11-27 08:46:53.994855315 +0000 UTC m=+0.086782310 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, version=17.1.12, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, tcib_managed=true, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:46:54 localhost podman[94067]: 2025-11-27 08:46:54.057604688 +0000 UTC m=+0.149531663 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=) Nov 27 03:46:54 localhost podman[94069]: 2025-11-27 08:46:54.068941582 +0000 UTC m=+0.157451605 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, tcib_managed=true, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, version=17.1.12, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute) Nov 27 03:46:54 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:46:54 localhost podman[94069]: 2025-11-27 08:46:54.127917474 +0000 UTC m=+0.216427497 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_id=tripleo_step4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, architecture=x86_64, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, io.openshift.expose-services=, vcs-type=git, version=17.1.12, tcib_managed=true) Nov 27 03:46:54 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:46:54 localhost podman[94068]: 2025-11-27 08:46:54.149696269 +0000 UTC m=+0.239967379 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, name=rhosp17/openstack-cron, io.buildah.version=1.41.4, config_id=tripleo_step4, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, container_name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, com.redhat.component=openstack-cron-container) Nov 27 03:46:54 localhost podman[94068]: 2025-11-27 08:46:54.185859669 +0000 UTC m=+0.276130739 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, distribution-scope=public, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, config_id=tripleo_step4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, release=1761123044, vendor=Red Hat, Inc., vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, url=https://www.redhat.com, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:46:54 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:46:54 localhost systemd[1]: tmp-crun.SZKbGl.mount: Deactivated successfully. Nov 27 03:46:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:46:56 localhost podman[94137]: 2025-11-27 08:46:56.961198203 +0000 UTC m=+0.058472680 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=nova_migration_target, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, vcs-type=git, io.openshift.expose-services=, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12) Nov 27 03:46:57 localhost podman[94137]: 2025-11-27 08:46:57.33600361 +0000 UTC m=+0.433278137 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, architecture=x86_64, batch=17.1_20251118.1, tcib_managed=true, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, config_id=tripleo_step4, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:46:57 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:46:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:46:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:46:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:46:59 localhost podman[94160]: 2025-11-27 08:46:59.998914636 +0000 UTC m=+0.092525363 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, architecture=x86_64, url=https://www.redhat.com, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, distribution-scope=public, tcib_managed=true, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, build-date=2025-11-19T00:14:25Z, vcs-type=git) Nov 27 03:47:00 localhost podman[94162]: 2025-11-27 08:47:00.066953071 +0000 UTC m=+0.154804254 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step5, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, architecture=x86_64, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:47:00 localhost podman[94160]: 2025-11-27 08:47:00.077352831 +0000 UTC m=+0.170963548 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_id=tripleo_step4, architecture=x86_64, distribution-scope=public, tcib_managed=true, release=1761123044, build-date=2025-11-19T00:14:25Z, vcs-type=git, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:47:00 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:47:00 localhost podman[94162]: 2025-11-27 08:47:00.155653181 +0000 UTC m=+0.243504394 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, release=1761123044, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, vcs-type=git, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:47:00 localhost podman[94161]: 2025-11-27 08:47:00.158163729 +0000 UTC m=+0.248127009 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, tcib_managed=true, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.41.4, release=1761123044, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, managed_by=tripleo_ansible, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, maintainer=OpenStack TripleO Team) Nov 27 03:47:00 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:47:00 localhost podman[94161]: 2025-11-27 08:47:00.242255815 +0000 UTC m=+0.332219095 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, architecture=x86_64, release=1761123044, build-date=2025-11-18T23:34:05Z, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, container_name=ovn_controller, vcs-type=git, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:47:00 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:47:04 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:47:04 localhost recover_tripleo_nova_virtqemud[94237]: 63639 Nov 27 03:47:04 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:47:04 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:47:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:47:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:47:08 localhost podman[94238]: 2025-11-27 08:47:08.996182837 +0000 UTC m=+0.085514846 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, managed_by=tripleo_ansible, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, url=https://www.redhat.com, vendor=Red Hat, Inc.) Nov 27 03:47:09 localhost podman[94238]: 2025-11-27 08:47:09.010905162 +0000 UTC m=+0.100237181 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, version=17.1.12, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-type=git, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd) Nov 27 03:47:09 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:47:09 localhost podman[94239]: 2025-11-27 08:47:09.058951931 +0000 UTC m=+0.142525926 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, tcib_managed=true, architecture=x86_64, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:47:09 localhost podman[94239]: 2025-11-27 08:47:09.072866084 +0000 UTC m=+0.156440069 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, container_name=iscsid, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, distribution-scope=public, io.buildah.version=1.41.4, vendor=Red Hat, Inc.) Nov 27 03:47:09 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:47:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:47:19 localhost systemd[1]: tmp-crun.yIRBkv.mount: Deactivated successfully. Nov 27 03:47:20 localhost podman[94276]: 2025-11-27 08:47:20.010947737 +0000 UTC m=+0.099494190 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, vendor=Red Hat, Inc., architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, tcib_managed=true, batch=17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:47:20 localhost podman[94276]: 2025-11-27 08:47:20.214940491 +0000 UTC m=+0.303486964 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, version=17.1.12, vcs-type=git, release=1761123044, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, distribution-scope=public, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:47:20 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:47:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:47:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:47:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:47:24 localhost podman[94304]: 2025-11-27 08:47:24.994241241 +0000 UTC m=+0.087781877 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vcs-type=git, io.buildah.version=1.41.4, tcib_managed=true, config_id=tripleo_step4, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:47:25 localhost systemd[1]: tmp-crun.VJbTj7.mount: Deactivated successfully. Nov 27 03:47:25 localhost podman[94304]: 2025-11-27 08:47:25.056266605 +0000 UTC m=+0.149807251 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, batch=17.1_20251118.1, url=https://www.redhat.com, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., distribution-scope=public, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_ipmi, vcs-type=git, config_id=tripleo_step4) Nov 27 03:47:25 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:47:25 localhost podman[94305]: 2025-11-27 08:47:25.059712288 +0000 UTC m=+0.150931231 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, name=rhosp17/openstack-cron, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, url=https://www.redhat.com, io.buildah.version=1.41.4, version=17.1.12, tcib_managed=true, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, vendor=Red Hat, Inc., container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible) Nov 27 03:47:25 localhost podman[94306]: 2025-11-27 08:47:25.115628698 +0000 UTC m=+0.203825170 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., config_id=tripleo_step4, container_name=ceilometer_agent_compute, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4) Nov 27 03:47:25 localhost podman[94305]: 2025-11-27 08:47:25.143965848 +0000 UTC m=+0.235184791 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, distribution-scope=public, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, url=https://www.redhat.com, release=1761123044, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:47:25 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:47:25 localhost podman[94306]: 2025-11-27 08:47:25.172614046 +0000 UTC m=+0.260810458 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, url=https://www.redhat.com, version=17.1.12, distribution-scope=public, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:47:25 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:47:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:47:27 localhost podman[94378]: 2025-11-27 08:47:27.992632349 +0000 UTC m=+0.091804984 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, distribution-scope=public, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, maintainer=OpenStack TripleO Team, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, container_name=nova_migration_target) Nov 27 03:47:28 localhost podman[94378]: 2025-11-27 08:47:28.363960632 +0000 UTC m=+0.463133217 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible) Nov 27 03:47:28 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:47:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:47:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:47:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:47:30 localhost systemd[1]: tmp-crun.cahTwa.mount: Deactivated successfully. Nov 27 03:47:31 localhost podman[94402]: 2025-11-27 08:47:30.993779071 +0000 UTC m=+0.087905219 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step4, release=1761123044, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, build-date=2025-11-19T00:14:25Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:47:31 localhost podman[94404]: 2025-11-27 08:47:31.060314016 +0000 UTC m=+0.146952433 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step5, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible) Nov 27 03:47:31 localhost podman[94403]: 2025-11-27 08:47:31.026612352 +0000 UTC m=+0.115025587 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, version=17.1.12, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, url=https://www.redhat.com, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, tcib_managed=true) Nov 27 03:47:31 localhost podman[94404]: 2025-11-27 08:47:31.08687876 +0000 UTC m=+0.173517137 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, vcs-type=git, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, architecture=x86_64, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, version=17.1.12, io.buildah.version=1.41.4, tcib_managed=true, config_id=tripleo_step5, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, container_name=nova_compute) Nov 27 03:47:31 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:47:31 localhost podman[94403]: 2025-11-27 08:47:31.109011053 +0000 UTC m=+0.197424338 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, release=1761123044, io.openshift.expose-services=, version=17.1.12, vcs-type=git, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:47:31 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Deactivated successfully. Nov 27 03:47:31 localhost podman[94402]: 2025-11-27 08:47:31.129117743 +0000 UTC m=+0.223243861 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, distribution-scope=public, io.openshift.expose-services=, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, batch=17.1_20251118.1) Nov 27 03:47:31 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:47:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:47:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:47:39 localhost podman[94587]: 2025-11-27 08:47:39.981846014 +0000 UTC m=+0.070478712 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vendor=Red Hat, Inc., managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, release=1761123044, vcs-type=git, url=https://www.redhat.com, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, name=rhosp17/openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:47:39 localhost podman[94587]: 2025-11-27 08:47:39.990231029 +0000 UTC m=+0.078863757 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, vendor=Red Hat, Inc., release=1761123044, url=https://www.redhat.com, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, vcs-type=git, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, tcib_managed=true, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3) Nov 27 03:47:40 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:47:40 localhost podman[94586]: 2025-11-27 08:47:40.042696766 +0000 UTC m=+0.131142389 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step3, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, vcs-type=git, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, managed_by=tripleo_ansible, release=1761123044, tcib_managed=true, container_name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:47:40 localhost podman[94586]: 2025-11-27 08:47:40.07675465 +0000 UTC m=+0.165200273 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, tcib_managed=true, name=rhosp17/openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044, vcs-type=git, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, io.openshift.expose-services=, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd) Nov 27 03:47:40 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:47:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:47:50 localhost systemd[1]: tmp-crun.l1qqSQ.mount: Deactivated successfully. Nov 27 03:47:50 localhost podman[94640]: 2025-11-27 08:47:50.99091216 +0000 UTC m=+0.089768330 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, batch=17.1_20251118.1, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.buildah.version=1.41.4) Nov 27 03:47:51 localhost podman[94640]: 2025-11-27 08:47:51.204209133 +0000 UTC m=+0.303065283 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, url=https://www.redhat.com, io.openshift.expose-services=, vcs-type=git, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, release=1761123044, name=rhosp17/openstack-qdrouterd) Nov 27 03:47:51 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 3600.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:47:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:47:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:47:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:47:55 localhost systemd[1]: tmp-crun.fzWjqt.mount: Deactivated successfully. Nov 27 03:47:56 localhost podman[94669]: 2025-11-27 08:47:56.004586037 +0000 UTC m=+0.099227713 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, architecture=x86_64, batch=17.1_20251118.1, distribution-scope=public, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, build-date=2025-11-19T00:12:45Z) Nov 27 03:47:56 localhost podman[94669]: 2025-11-27 08:47:56.033883734 +0000 UTC m=+0.128525390 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, tcib_managed=true, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, maintainer=OpenStack TripleO Team, version=17.1.12, build-date=2025-11-19T00:12:45Z, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:47:56 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:47:56 localhost podman[94670]: 2025-11-27 08:47:56.051756883 +0000 UTC m=+0.142900165 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, distribution-scope=public, vendor=Red Hat, Inc., batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, version=17.1.12, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_step4, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:47:56 localhost podman[94670]: 2025-11-27 08:47:56.05985436 +0000 UTC m=+0.150997652 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, batch=17.1_20251118.1, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.openshift.expose-services=, com.redhat.component=openstack-cron-container, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4, container_name=logrotate_crond, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, maintainer=OpenStack TripleO Team) Nov 27 03:47:56 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:47:56 localhost podman[94671]: 2025-11-27 08:47:56.112541244 +0000 UTC m=+0.197367077 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, release=1761123044, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-type=git) Nov 27 03:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 3600.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:47:56 localhost podman[94671]: 2025-11-27 08:47:56.167887509 +0000 UTC m=+0.252713332 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, config_id=tripleo_step4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, batch=17.1_20251118.1, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, container_name=ceilometer_agent_compute, io.openshift.expose-services=) Nov 27 03:47:56 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:47:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:47:58 localhost podman[94741]: 2025-11-27 08:47:58.983462433 +0000 UTC m=+0.081991642 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, distribution-scope=public, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, architecture=x86_64, batch=17.1_20251118.1, url=https://www.redhat.com, version=17.1.12, vendor=Red Hat, Inc., vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, container_name=nova_migration_target, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:47:59 localhost podman[94741]: 2025-11-27 08:47:59.336948587 +0000 UTC m=+0.435477796 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, version=17.1.12, vcs-type=git, container_name=nova_migration_target, batch=17.1_20251118.1, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, architecture=x86_64, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:47:59 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:48:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:48:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:48:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:48:01 localhost systemd[1]: tmp-crun.cZ5RA2.mount: Deactivated successfully. Nov 27 03:48:02 localhost podman[94765]: 2025-11-27 08:48:02.002514114 +0000 UTC m=+0.094309650 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, version=17.1.12, config_id=tripleo_step4, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z) Nov 27 03:48:02 localhost podman[94766]: 2025-11-27 08:48:02.031707228 +0000 UTC m=+0.114295497 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step5, maintainer=OpenStack TripleO Team, tcib_managed=true, url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:48:02 localhost podman[94766]: 2025-11-27 08:48:02.05862684 +0000 UTC m=+0.141215129 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, container_name=nova_compute, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, architecture=x86_64, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, vcs-type=git, config_id=tripleo_step5, io.openshift.expose-services=) Nov 27 03:48:02 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:48:02 localhost podman[94765]: 2025-11-27 08:48:02.083900488 +0000 UTC m=+0.175696054 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, release=1761123044, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z) Nov 27 03:48:02 localhost podman[94765]: unhealthy Nov 27 03:48:02 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:48:02 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:48:02 localhost podman[94764]: 2025-11-27 08:48:02.147081144 +0000 UTC m=+0.241997554 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vendor=Red Hat, Inc., release=1761123044, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, version=17.1.12, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, url=https://www.redhat.com, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, distribution-scope=public, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team) Nov 27 03:48:02 localhost podman[94764]: 2025-11-27 08:48:02.21701351 +0000 UTC m=+0.311929920 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, container_name=ovn_metadata_agent, vcs-type=git, url=https://www.redhat.com, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:48:02 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:48:02 localhost systemd[1]: tmp-crun.PZtuLW.mount: Deactivated successfully. Nov 27 03:48:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:48:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:48:10 localhost systemd[1]: tmp-crun.OCtpvC.mount: Deactivated successfully. Nov 27 03:48:10 localhost podman[94843]: 2025-11-27 08:48:10.996721063 +0000 UTC m=+0.087488688 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, url=https://www.redhat.com, io.openshift.expose-services=, io.buildah.version=1.41.4, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, architecture=x86_64, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, vendor=Red Hat, Inc., distribution-scope=public, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:48:11 localhost systemd[1]: tmp-crun.PVORy7.mount: Deactivated successfully. Nov 27 03:48:11 localhost podman[94844]: 2025-11-27 08:48:11.040642712 +0000 UTC m=+0.128806467 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, distribution-scope=public, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:48:11 localhost podman[94843]: 2025-11-27 08:48:11.059388574 +0000 UTC m=+0.150156259 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-collectd-container, container_name=collectd, vcs-type=git, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, vendor=Red Hat, Inc., config_id=tripleo_step3, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, url=https://www.redhat.com, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 03:48:11 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:48:11 localhost podman[94844]: 2025-11-27 08:48:11.077002368 +0000 UTC m=+0.165166153 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, release=1761123044, distribution-scope=public, version=17.1.12, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.buildah.version=1.41.4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, container_name=iscsid, url=https://www.redhat.com) Nov 27 03:48:11 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:48:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:48:21 localhost systemd[1]: tmp-crun.A06ul9.mount: Deactivated successfully. Nov 27 03:48:21 localhost podman[94881]: 2025-11-27 08:48:21.994479317 +0000 UTC m=+0.094026194 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, release=1761123044, vendor=Red Hat, Inc., config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, vcs-type=git, name=rhosp17/openstack-qdrouterd, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd) Nov 27 03:48:22 localhost podman[94881]: 2025-11-27 08:48:22.214112809 +0000 UTC m=+0.313659646 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-type=git, managed_by=tripleo_ansible, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:49:46Z, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, name=rhosp17/openstack-qdrouterd) Nov 27 03:48:22 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:48:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:48:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:48:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:48:26 localhost systemd[1]: tmp-crun.KKydvo.mount: Deactivated successfully. Nov 27 03:48:26 localhost podman[94909]: 2025-11-27 08:48:26.991980631 +0000 UTC m=+0.091544497 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vendor=Red Hat, Inc., io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, architecture=x86_64, build-date=2025-11-19T00:12:45Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true) Nov 27 03:48:27 localhost podman[94909]: 2025-11-27 08:48:27.025052529 +0000 UTC m=+0.124616375 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, version=17.1.12, tcib_managed=true, architecture=x86_64, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4) Nov 27 03:48:27 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:48:27 localhost podman[94911]: 2025-11-27 08:48:27.043694248 +0000 UTC m=+0.140517301 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, version=17.1.12, container_name=ceilometer_agent_compute, vcs-type=git, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, release=1761123044) Nov 27 03:48:27 localhost podman[94910]: 2025-11-27 08:48:27.080870825 +0000 UTC m=+0.178759416 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.buildah.version=1.41.4, vendor=Red Hat, Inc., distribution-scope=public, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, config_id=tripleo_step4, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:48:27 localhost podman[94910]: 2025-11-27 08:48:27.093758632 +0000 UTC m=+0.191647213 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, version=17.1.12, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, container_name=logrotate_crond, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, vendor=Red Hat, Inc., managed_by=tripleo_ansible, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team) Nov 27 03:48:27 localhost podman[94911]: 2025-11-27 08:48:27.099974018 +0000 UTC m=+0.196797031 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, release=1761123044, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, version=17.1.12, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:48:27 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:48:27 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:48:27 localhost systemd[1]: tmp-crun.bsVwae.mount: Deactivated successfully. Nov 27 03:48:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:48:29 localhost podman[94982]: 2025-11-27 08:48:29.978809469 +0000 UTC m=+0.078061006 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., distribution-scope=public, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:48:30 localhost podman[94982]: 2025-11-27 08:48:30.345926889 +0000 UTC m=+0.445178426 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, distribution-scope=public, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, version=17.1.12, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container) Nov 27 03:48:30 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:48:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:48:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:48:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:48:32 localhost podman[95006]: 2025-11-27 08:48:32.985646644 +0000 UTC m=+0.079629717 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, vcs-type=git, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible) Nov 27 03:48:33 localhost podman[95006]: 2025-11-27 08:48:33.032976234 +0000 UTC m=+0.126959317 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, vcs-type=git, version=17.1.12, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:48:33 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Deactivated successfully. Nov 27 03:48:33 localhost podman[95007]: 2025-11-27 08:48:33.053494205 +0000 UTC m=+0.145586917 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, architecture=x86_64, batch=17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, release=1761123044, vendor=Red Hat, Inc., io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, vcs-type=git, maintainer=OpenStack TripleO Team) Nov 27 03:48:33 localhost podman[95008]: 2025-11-27 08:48:33.10736614 +0000 UTC m=+0.195590819 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, tcib_managed=true, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., vcs-type=git, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, version=17.1.12, build-date=2025-11-19T00:36:58Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:48:33 localhost podman[95007]: 2025-11-27 08:48:33.130203933 +0000 UTC m=+0.222296645 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, release=1761123044, com.redhat.component=openstack-ovn-controller-container, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.buildah.version=1.41.4, batch=17.1_20251118.1, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 03:48:33 localhost podman[95007]: unhealthy Nov 27 03:48:33 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:48:33 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:48:33 localhost podman[95008]: 2025-11-27 08:48:33.162943071 +0000 UTC m=+0.251167740 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, version=17.1.12, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, architecture=x86_64, io.openshift.expose-services=, io.buildah.version=1.41.4, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, vcs-type=git) Nov 27 03:48:33 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:48:33 localhost systemd[1]: tmp-crun.5VPszo.mount: Deactivated successfully. Nov 27 03:48:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:48:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:48:41 localhost podman[95160]: 2025-11-27 08:48:41.752714238 +0000 UTC m=+0.088395843 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, url=https://www.redhat.com, architecture=x86_64, io.openshift.expose-services=, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, version=17.1.12, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:48:41 localhost systemd[1]: tmp-crun.oJhrKC.mount: Deactivated successfully. Nov 27 03:48:41 localhost podman[95159]: 2025-11-27 08:48:41.777123573 +0000 UTC m=+0.114107553 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, architecture=x86_64, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, container_name=collectd, io.openshift.expose-services=, version=17.1.12, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:48:41 localhost podman[95159]: 2025-11-27 08:48:41.790951733 +0000 UTC m=+0.127935683 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, vcs-type=git, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, release=1761123044, config_id=tripleo_step3, distribution-scope=public, vendor=Red Hat, Inc., io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:48:41 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:48:41 localhost podman[95160]: 2025-11-27 08:48:41.841978993 +0000 UTC m=+0.177660528 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.buildah.version=1.41.4, name=rhosp17/openstack-iscsid, release=1761123044, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3) Nov 27 03:48:41 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:48:42 localhost podman[95253]: Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.376151114 +0000 UTC m=+0.080292615 container create 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., architecture=x86_64, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, version=7, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, distribution-scope=public, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-type=git, io.openshift.expose-services=, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , ceph=True, build-date=2025-09-24T08:57:55) Nov 27 03:48:42 localhost systemd[1]: Started libpod-conmon-37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707.scope. Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.34247039 +0000 UTC m=+0.046611951 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 03:48:42 localhost systemd[1]: Started libcrun container. Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.461462944 +0000 UTC m=+0.165604455 container init 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, release=553, distribution-scope=public, vendor=Red Hat, Inc., architecture=x86_64, version=7, vcs-type=git) Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.473003163 +0000 UTC m=+0.177144674 container start 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, version=7, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, maintainer=Guillaume Abrioux , release=553, CEPH_POINT_RELEASE=, architecture=x86_64, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, vcs-type=git) Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.473330192 +0000 UTC m=+0.177471703 container attach 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, name=rhceph, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, release=553, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, ceph=True, vendor=Red Hat, Inc., RELEASE=main) Nov 27 03:48:42 localhost elegant_allen[95269]: 167 167 Nov 27 03:48:42 localhost systemd[1]: libpod-37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707.scope: Deactivated successfully. Nov 27 03:48:42 localhost podman[95253]: 2025-11-27 08:48:42.478336336 +0000 UTC m=+0.182477867 container died 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., vcs-type=git, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, RELEASE=main, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12) Nov 27 03:48:42 localhost podman[95274]: 2025-11-27 08:48:42.579621004 +0000 UTC m=+0.088994139 container remove 37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elegant_allen, CEPH_POINT_RELEASE=, release=553, ceph=True, distribution-scope=public, architecture=x86_64, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.buildah.version=1.33.12, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git) Nov 27 03:48:42 localhost systemd[1]: libpod-conmon-37429ed74cc8729f9fdb1118ca0997c90fe36b46820234c1ad032b5b55213707.scope: Deactivated successfully. Nov 27 03:48:42 localhost systemd[1]: var-lib-containers-storage-overlay-2eee07098a05eb50bcb9571afa60f99575cc592b1c6a89a3a899cf9d8515b5de-merged.mount: Deactivated successfully. Nov 27 03:48:42 localhost podman[95297]: Nov 27 03:48:42 localhost podman[95297]: 2025-11-27 08:48:42.830812884 +0000 UTC m=+0.083182344 container create 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, version=7, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., distribution-scope=public, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, ceph=True, CEPH_POINT_RELEASE=, architecture=x86_64, maintainer=Guillaume Abrioux , vcs-type=git, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, release=553, build-date=2025-09-24T08:57:55) Nov 27 03:48:42 localhost systemd[1]: Started libpod-conmon-63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806.scope. Nov 27 03:48:42 localhost podman[95297]: 2025-11-27 08:48:42.801127997 +0000 UTC m=+0.053497457 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 03:48:42 localhost systemd[1]: Started libcrun container. Nov 27 03:48:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/bbb039c5f1549b4bcda3693044320be68b21c52358197e21f7eee3f23bb41660/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 03:48:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/bbb039c5f1549b4bcda3693044320be68b21c52358197e21f7eee3f23bb41660/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 03:48:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/bbb039c5f1549b4bcda3693044320be68b21c52358197e21f7eee3f23bb41660/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 03:48:42 localhost podman[95297]: 2025-11-27 08:48:42.917436528 +0000 UTC m=+0.169805998 container init 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, name=rhceph, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, io.buildah.version=1.33.12, architecture=x86_64, ceph=True, build-date=2025-09-24T08:57:55, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, distribution-scope=public, GIT_CLEAN=True, description=Red Hat Ceph Storage 7) Nov 27 03:48:42 localhost podman[95297]: 2025-11-27 08:48:42.930286462 +0000 UTC m=+0.182655922 container start 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, GIT_CLEAN=True, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , distribution-scope=public, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, io.buildah.version=1.33.12, vendor=Red Hat, Inc., release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 03:48:42 localhost podman[95297]: 2025-11-27 08:48:42.930616911 +0000 UTC m=+0.182986421 container attach 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, vcs-type=git, GIT_BRANCH=main, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, RELEASE=main, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, release=553, build-date=2025-09-24T08:57:55, name=rhceph) Nov 27 03:48:43 localhost elastic_payne[95313]: [ Nov 27 03:48:43 localhost elastic_payne[95313]: { Nov 27 03:48:43 localhost elastic_payne[95313]: "available": false, Nov 27 03:48:43 localhost elastic_payne[95313]: "ceph_device": false, Nov 27 03:48:43 localhost elastic_payne[95313]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 03:48:43 localhost elastic_payne[95313]: "lsm_data": {}, Nov 27 03:48:43 localhost elastic_payne[95313]: "lvs": [], Nov 27 03:48:43 localhost elastic_payne[95313]: "path": "/dev/sr0", Nov 27 03:48:43 localhost elastic_payne[95313]: "rejected_reasons": [ Nov 27 03:48:43 localhost elastic_payne[95313]: "Insufficient space (<5GB)", Nov 27 03:48:43 localhost elastic_payne[95313]: "Has a FileSystem" Nov 27 03:48:43 localhost elastic_payne[95313]: ], Nov 27 03:48:43 localhost elastic_payne[95313]: "sys_api": { Nov 27 03:48:43 localhost elastic_payne[95313]: "actuators": null, Nov 27 03:48:43 localhost elastic_payne[95313]: "device_nodes": "sr0", Nov 27 03:48:43 localhost elastic_payne[95313]: "human_readable_size": "482.00 KB", Nov 27 03:48:43 localhost elastic_payne[95313]: "id_bus": "ata", Nov 27 03:48:43 localhost elastic_payne[95313]: "model": "QEMU DVD-ROM", Nov 27 03:48:43 localhost elastic_payne[95313]: "nr_requests": "2", Nov 27 03:48:43 localhost elastic_payne[95313]: "partitions": {}, Nov 27 03:48:43 localhost elastic_payne[95313]: "path": "/dev/sr0", Nov 27 03:48:43 localhost elastic_payne[95313]: "removable": "1", Nov 27 03:48:43 localhost elastic_payne[95313]: "rev": "2.5+", Nov 27 03:48:43 localhost elastic_payne[95313]: "ro": "0", Nov 27 03:48:43 localhost elastic_payne[95313]: "rotational": "1", Nov 27 03:48:43 localhost elastic_payne[95313]: "sas_address": "", Nov 27 03:48:43 localhost elastic_payne[95313]: "sas_device_handle": "", Nov 27 03:48:43 localhost elastic_payne[95313]: "scheduler_mode": "mq-deadline", Nov 27 03:48:43 localhost elastic_payne[95313]: "sectors": 0, Nov 27 03:48:43 localhost elastic_payne[95313]: "sectorsize": "2048", Nov 27 03:48:43 localhost elastic_payne[95313]: "size": 493568.0, Nov 27 03:48:43 localhost elastic_payne[95313]: "support_discard": "0", Nov 27 03:48:43 localhost elastic_payne[95313]: "type": "disk", Nov 27 03:48:43 localhost elastic_payne[95313]: "vendor": "QEMU" Nov 27 03:48:43 localhost elastic_payne[95313]: } Nov 27 03:48:43 localhost elastic_payne[95313]: } Nov 27 03:48:43 localhost elastic_payne[95313]: ] Nov 27 03:48:43 localhost systemd[1]: libpod-63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806.scope: Deactivated successfully. Nov 27 03:48:43 localhost podman[95297]: 2025-11-27 08:48:43.986992183 +0000 UTC m=+1.239361683 container died 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, vendor=Red Hat, Inc., RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, GIT_CLEAN=True, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12) Nov 27 03:48:43 localhost systemd[1]: libpod-63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806.scope: Consumed 1.081s CPU time. Nov 27 03:48:44 localhost systemd[1]: var-lib-containers-storage-overlay-bbb039c5f1549b4bcda3693044320be68b21c52358197e21f7eee3f23bb41660-merged.mount: Deactivated successfully. Nov 27 03:48:44 localhost podman[97170]: 2025-11-27 08:48:44.093208493 +0000 UTC m=+0.097192538 container remove 63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=elastic_payne, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, maintainer=Guillaume Abrioux , name=rhceph, vendor=Red Hat, Inc., RELEASE=main, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, com.redhat.component=rhceph-container, distribution-scope=public, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 03:48:44 localhost systemd[1]: libpod-conmon-63fa1938d8882525b05b10a3abf0c2ddcaaac5ef9bc44d47db30323c8f929806.scope: Deactivated successfully. Nov 27 03:48:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:48:52 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:48:52 localhost recover_tripleo_nova_virtqemud[97201]: 63639 Nov 27 03:48:52 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:48:52 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:48:52 localhost podman[97199]: 2025-11-27 08:48:52.994781356 +0000 UTC m=+0.088821114 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, batch=17.1_20251118.1, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, container_name=metrics_qdr, vendor=Red Hat, Inc., distribution-scope=public, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:48:53 localhost podman[97199]: 2025-11-27 08:48:53.227066539 +0000 UTC m=+0.321106277 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true) Nov 27 03:48:53 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:48:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:48:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:48:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:48:58 localhost podman[97232]: 2025-11-27 08:48:57.996716999 +0000 UTC m=+0.091494925 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-type=git, version=17.1.12, config_id=tripleo_step4, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, io.buildah.version=1.41.4, container_name=logrotate_crond, io.openshift.expose-services=, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, tcib_managed=true) Nov 27 03:48:58 localhost podman[97232]: 2025-11-27 08:48:58.008953078 +0000 UTC m=+0.103731064 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, release=1761123044, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, config_id=tripleo_step4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, container_name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, architecture=x86_64, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:48:58 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:48:58 localhost systemd[1]: tmp-crun.1c1L3V.mount: Deactivated successfully. Nov 27 03:48:58 localhost podman[97231]: 2025-11-27 08:48:58.101121421 +0000 UTC m=+0.197608493 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., distribution-scope=public, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:48:58 localhost podman[97231]: 2025-11-27 08:48:58.13201682 +0000 UTC m=+0.228503912 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, release=1761123044, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, architecture=x86_64, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:48:58 localhost podman[97233]: 2025-11-27 08:48:58.151974496 +0000 UTC m=+0.245151929 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.expose-services=, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_id=tripleo_step4, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, vendor=Red Hat, Inc.) Nov 27 03:48:58 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:48:58 localhost podman[97233]: 2025-11-27 08:48:58.212003196 +0000 UTC m=+0.305180639 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, architecture=x86_64, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, release=1761123044, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, maintainer=OpenStack TripleO Team) Nov 27 03:48:58 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:49:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:49:00 localhost podman[97301]: 2025-11-27 08:49:00.99407303 +0000 UTC m=+0.090246492 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, container_name=nova_migration_target, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4) Nov 27 03:49:01 localhost podman[97301]: 2025-11-27 08:49:01.374796885 +0000 UTC m=+0.470970347 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_id=tripleo_step4, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, release=1761123044, distribution-scope=public, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.openshift.expose-services=) Nov 27 03:49:01 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:49:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:49:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:49:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:49:03 localhost podman[97324]: 2025-11-27 08:49:03.994175523 +0000 UTC m=+0.087747645 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.openshift.expose-services=, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, container_name=ovn_metadata_agent, batch=17.1_20251118.1, vcs-type=git, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, release=1761123044, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}) Nov 27 03:49:04 localhost podman[97324]: 2025-11-27 08:49:04.007641735 +0000 UTC m=+0.101213857 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, release=1761123044, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., distribution-scope=public, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true) Nov 27 03:49:04 localhost podman[97324]: unhealthy Nov 27 03:49:04 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:49:04 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:49:04 localhost podman[97325]: 2025-11-27 08:49:04.104287928 +0000 UTC m=+0.192311831 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, version=17.1.12, container_name=ovn_controller, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, distribution-scope=public, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:49:04 localhost podman[97325]: 2025-11-27 08:49:04.152591504 +0000 UTC m=+0.240615397 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, tcib_managed=true, distribution-scope=public, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64) Nov 27 03:49:04 localhost podman[97325]: unhealthy Nov 27 03:49:04 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:49:04 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:49:04 localhost podman[97326]: 2025-11-27 08:49:04.157733642 +0000 UTC m=+0.244849671 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, vcs-type=git, release=1761123044, distribution-scope=public, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, io.buildah.version=1.41.4, version=17.1.12, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, config_id=tripleo_step5, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:49:04 localhost podman[97326]: 2025-11-27 08:49:04.237264166 +0000 UTC m=+0.324380195 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, config_id=tripleo_step5, url=https://www.redhat.com, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:49:04 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:49:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:49:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:49:11 localhost podman[97390]: 2025-11-27 08:49:11.981251758 +0000 UTC m=+0.080914281 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, batch=17.1_20251118.1, distribution-scope=public, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, vcs-type=git, architecture=x86_64, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:49:11 localhost podman[97390]: 2025-11-27 08:49:11.991010091 +0000 UTC m=+0.090672604 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, container_name=collectd, tcib_managed=true, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., version=17.1.12, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:49:12 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:49:12 localhost podman[97391]: 2025-11-27 08:49:12.078093607 +0000 UTC m=+0.176344252 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, architecture=x86_64, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, managed_by=tripleo_ansible, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, url=https://www.redhat.com, distribution-scope=public, release=1761123044, vcs-type=git) Nov 27 03:49:12 localhost podman[97391]: 2025-11-27 08:49:12.086823101 +0000 UTC m=+0.185073716 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, architecture=x86_64, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, release=1761123044, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team) Nov 27 03:49:12 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:49:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:49:23 localhost podman[97428]: 2025-11-27 08:49:23.982205159 +0000 UTC m=+0.081351394 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., batch=17.1_20251118.1, tcib_managed=true, architecture=x86_64, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_id=tripleo_step1) Nov 27 03:49:24 localhost podman[97428]: 2025-11-27 08:49:24.179480912 +0000 UTC m=+0.278627117 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, batch=17.1_20251118.1, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, architecture=x86_64, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, version=17.1.12, build-date=2025-11-18T22:49:46Z) Nov 27 03:49:24 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:49:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:49:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:49:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:49:28 localhost systemd[1]: tmp-crun.rHcFf3.mount: Deactivated successfully. Nov 27 03:49:28 localhost podman[97457]: 2025-11-27 08:49:28.994079599 +0000 UTC m=+0.084170869 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, com.redhat.component=openstack-cron-container, distribution-scope=public, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_id=tripleo_step4, io.buildah.version=1.41.4, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, release=1761123044, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12) Nov 27 03:49:29 localhost podman[97457]: 2025-11-27 08:49:29.031994837 +0000 UTC m=+0.122086097 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, version=17.1.12, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, config_id=tripleo_step4, build-date=2025-11-18T22:49:32Z, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=logrotate_crond, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044) Nov 27 03:49:29 localhost systemd[1]: tmp-crun.hlehRi.mount: Deactivated successfully. Nov 27 03:49:29 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:49:29 localhost podman[97456]: 2025-11-27 08:49:29.050821892 +0000 UTC m=+0.140548623 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, build-date=2025-11-19T00:12:45Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, config_id=tripleo_step4, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:49:29 localhost podman[97458]: 2025-11-27 08:49:29.094041471 +0000 UTC m=+0.182253131 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, tcib_managed=true, build-date=2025-11-19T00:11:48Z, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, io.openshift.expose-services=, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, version=17.1.12, vendor=Red Hat, Inc., url=https://www.redhat.com, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:49:29 localhost podman[97456]: 2025-11-27 08:49:29.103808314 +0000 UTC m=+0.193534995 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:12:45Z, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, tcib_managed=true, io.openshift.expose-services=, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4) Nov 27 03:49:29 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:49:29 localhost podman[97458]: 2025-11-27 08:49:29.126933484 +0000 UTC m=+0.215145194 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:49:29 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:49:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:49:31 localhost podman[97525]: 2025-11-27 08:49:31.976029576 +0000 UTC m=+0.073002389 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, vendor=Red Hat, Inc., io.buildah.version=1.41.4, release=1761123044, container_name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:49:32 localhost podman[97525]: 2025-11-27 08:49:32.370022307 +0000 UTC m=+0.466995140 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, vcs-type=git, url=https://www.redhat.com, architecture=x86_64, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4) Nov 27 03:49:32 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:49:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:49:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:49:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:49:34 localhost systemd[1]: tmp-crun.f5TbXh.mount: Deactivated successfully. Nov 27 03:49:34 localhost podman[97548]: 2025-11-27 08:49:34.994978166 +0000 UTC m=+0.089051410 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, tcib_managed=true, build-date=2025-11-19T00:14:25Z, distribution-scope=public, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, io.openshift.expose-services=, managed_by=tripleo_ansible, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent) Nov 27 03:49:35 localhost podman[97548]: 2025-11-27 08:49:35.009138556 +0000 UTC m=+0.103211800 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, architecture=x86_64, io.openshift.expose-services=, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, distribution-scope=public, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:49:35 localhost podman[97548]: unhealthy Nov 27 03:49:35 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:49:35 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:49:35 localhost podman[97550]: 2025-11-27 08:49:35.061760217 +0000 UTC m=+0.149302337 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step5, release=1761123044, container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, url=https://www.redhat.com, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:49:35 localhost podman[97550]: 2025-11-27 08:49:35.088572217 +0000 UTC m=+0.176114307 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.openshift.expose-services=, vcs-type=git, version=17.1.12, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step5, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, build-date=2025-11-19T00:36:58Z, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., container_name=nova_compute) Nov 27 03:49:35 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:49:35 localhost podman[97549]: 2025-11-27 08:49:35.146942593 +0000 UTC m=+0.237431512 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., distribution-scope=public) Nov 27 03:49:35 localhost podman[97549]: 2025-11-27 08:49:35.163706083 +0000 UTC m=+0.254195052 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, version=17.1.12, config_id=tripleo_step4, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true) Nov 27 03:49:35 localhost podman[97549]: unhealthy Nov 27 03:49:35 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:49:35 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:49:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:49:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:49:42 localhost podman[97615]: 2025-11-27 08:49:42.979679958 +0000 UTC m=+0.078392645 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, container_name=collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, release=1761123044, com.redhat.component=openstack-collectd-container, version=17.1.12, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:51:28Z, maintainer=OpenStack TripleO Team, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, managed_by=tripleo_ansible, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git) Nov 27 03:49:42 localhost podman[97615]: 2025-11-27 08:49:42.990432356 +0000 UTC m=+0.089145043 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, container_name=collectd, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true) Nov 27 03:49:43 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:49:43 localhost podman[97616]: 2025-11-27 08:49:43.034848367 +0000 UTC m=+0.130198554 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.component=openstack-iscsid-container, container_name=iscsid, config_id=tripleo_step3, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, version=17.1.12, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:49:43 localhost podman[97616]: 2025-11-27 08:49:43.043127949 +0000 UTC m=+0.138478126 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, name=rhosp17/openstack-iscsid, distribution-scope=public, managed_by=tripleo_ansible) Nov 27 03:49:43 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:49:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:49:55 localhost podman[97782]: 2025-11-27 08:49:55.003219103 +0000 UTC m=+0.098099323 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, tcib_managed=true, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, build-date=2025-11-18T22:49:46Z, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, version=17.1.12, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=) Nov 27 03:49:55 localhost podman[97782]: 2025-11-27 08:49:55.22111088 +0000 UTC m=+0.315991080 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, distribution-scope=public, version=17.1.12) Nov 27 03:49:55 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:49:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:49:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:49:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:49:59 localhost systemd[1]: tmp-crun.EIRvLI.mount: Deactivated successfully. Nov 27 03:50:00 localhost podman[97812]: 2025-11-27 08:50:00.006103492 +0000 UTC m=+0.101230897 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, vendor=Red Hat, Inc., config_id=tripleo_step4, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-ipmi, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi) Nov 27 03:50:00 localhost systemd[1]: tmp-crun.MOi3r7.mount: Deactivated successfully. Nov 27 03:50:00 localhost podman[97813]: 2025-11-27 08:50:00.058729133 +0000 UTC m=+0.147266962 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, url=https://www.redhat.com, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, managed_by=tripleo_ansible, com.redhat.component=openstack-cron-container, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=logrotate_crond, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64) Nov 27 03:50:00 localhost podman[97812]: 2025-11-27 08:50:00.066365528 +0000 UTC m=+0.161492933 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, distribution-scope=public, version=17.1.12, release=1761123044, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-type=git, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:50:00 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:50:00 localhost podman[97813]: 2025-11-27 08:50:00.091636386 +0000 UTC m=+0.180174195 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, tcib_managed=true, io.openshift.expose-services=, container_name=logrotate_crond, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, com.redhat.component=openstack-cron-container, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, build-date=2025-11-18T22:49:32Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-type=git, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.buildah.version=1.41.4, release=1761123044, batch=17.1_20251118.1) Nov 27 03:50:00 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:50:00 localhost podman[97814]: 2025-11-27 08:50:00.158470459 +0000 UTC m=+0.246490774 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com) Nov 27 03:50:00 localhost podman[97814]: 2025-11-27 08:50:00.18791316 +0000 UTC m=+0.275933475 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step4, version=17.1.12, name=rhosp17/openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, vcs-type=git, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z) Nov 27 03:50:00 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:50:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:50:03 localhost systemd[1]: tmp-crun.5XTeD5.mount: Deactivated successfully. Nov 27 03:50:03 localhost podman[97886]: 2025-11-27 08:50:03.011576049 +0000 UTC m=+0.101425763 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, vcs-type=git, distribution-scope=public, container_name=nova_migration_target, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=) Nov 27 03:50:03 localhost podman[97886]: 2025-11-27 08:50:03.410167703 +0000 UTC m=+0.500017497 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, version=17.1.12) Nov 27 03:50:03 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:50:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:50:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:50:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:50:05 localhost systemd[1]: tmp-crun.3LJxb2.mount: Deactivated successfully. Nov 27 03:50:05 localhost podman[97909]: 2025-11-27 08:50:05.987442042 +0000 UTC m=+0.080557823 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, container_name=ovn_controller, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, distribution-scope=public, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, architecture=x86_64, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1) Nov 27 03:50:06 localhost podman[97909]: 2025-11-27 08:50:06.03098185 +0000 UTC m=+0.124097561 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, distribution-scope=public, vcs-type=git, config_id=tripleo_step4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vendor=Red Hat, Inc., url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:50:06 localhost systemd[1]: tmp-crun.AiHa38.mount: Deactivated successfully. Nov 27 03:50:06 localhost podman[97909]: unhealthy Nov 27 03:50:06 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:50:06 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:50:06 localhost podman[97908]: 2025-11-27 08:50:06.047407381 +0000 UTC m=+0.141509558 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=healthy, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, managed_by=tripleo_ansible, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, batch=17.1_20251118.1) Nov 27 03:50:06 localhost podman[97908]: 2025-11-27 08:50:06.05595864 +0000 UTC m=+0.150060787 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public) Nov 27 03:50:06 localhost podman[97908]: unhealthy Nov 27 03:50:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:50:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:50:06 localhost podman[97910]: 2025-11-27 08:50:06.144419424 +0000 UTC m=+0.233749953 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, distribution-scope=public, tcib_managed=true, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, container_name=nova_compute, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-nova-compute-container, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:50:06 localhost podman[97910]: 2025-11-27 08:50:06.171057758 +0000 UTC m=+0.260388247 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, distribution-scope=public, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, config_id=tripleo_step5, vcs-type=git, batch=17.1_20251118.1, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, version=17.1.12, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute) Nov 27 03:50:06 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:50:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:50:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:50:13 localhost systemd[1]: tmp-crun.A3rLVl.mount: Deactivated successfully. Nov 27 03:50:14 localhost podman[97974]: 2025-11-27 08:50:14.00137633 +0000 UTC m=+0.092701879 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, build-date=2025-11-18T22:51:28Z, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, architecture=x86_64, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, version=17.1.12, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3) Nov 27 03:50:14 localhost podman[97974]: 2025-11-27 08:50:14.016424623 +0000 UTC m=+0.107750192 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_id=tripleo_step3, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, distribution-scope=public, name=rhosp17/openstack-collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, tcib_managed=true, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, managed_by=tripleo_ansible) Nov 27 03:50:14 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:50:14 localhost podman[97975]: 2025-11-27 08:50:14.085192298 +0000 UTC m=+0.174343029 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vcs-type=git, managed_by=tripleo_ansible, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, version=17.1.12, distribution-scope=public, name=rhosp17/openstack-iscsid, release=1761123044, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:50:14 localhost podman[97975]: 2025-11-27 08:50:14.093898322 +0000 UTC m=+0.183049013 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, release=1761123044, io.buildah.version=1.41.4, url=https://www.redhat.com, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, batch=17.1_20251118.1, config_id=tripleo_step3, tcib_managed=true, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, container_name=iscsid, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:50:14 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:50:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:50:25 localhost podman[98013]: 2025-11-27 08:50:25.994799576 +0000 UTC m=+0.091914127 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, config_id=tripleo_step1, distribution-scope=public, vcs-type=git, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:50:26 localhost podman[98013]: 2025-11-27 08:50:26.212163847 +0000 UTC m=+0.309278398 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, version=17.1.12, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:50:26 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:50:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:50:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:50:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:50:30 localhost systemd[1]: tmp-crun.Vw1Eaa.mount: Deactivated successfully. Nov 27 03:50:31 localhost podman[98040]: 2025-11-27 08:50:31.002098373 +0000 UTC m=+0.093271463 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, vcs-type=git, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, url=https://www.redhat.com, managed_by=tripleo_ansible, config_id=tripleo_step4, tcib_managed=true) Nov 27 03:50:31 localhost podman[98041]: 2025-11-27 08:50:31.054671974 +0000 UTC m=+0.141958040 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., version=17.1.12, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, io.buildah.version=1.41.4, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond) Nov 27 03:50:31 localhost podman[98040]: 2025-11-27 08:50:31.06124335 +0000 UTC m=+0.152416450 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com) Nov 27 03:50:31 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:50:31 localhost podman[98041]: 2025-11-27 08:50:31.094191844 +0000 UTC m=+0.181477930 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, com.redhat.component=openstack-cron-container, architecture=x86_64, batch=17.1_20251118.1, name=rhosp17/openstack-cron, version=17.1.12, url=https://www.redhat.com, managed_by=tripleo_ansible, config_id=tripleo_step4, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, build-date=2025-11-18T22:49:32Z, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, container_name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:50:31 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:50:31 localhost podman[98042]: 2025-11-27 08:50:31.116949464 +0000 UTC m=+0.197760846 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, container_name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, batch=17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, vendor=Red Hat, Inc.) Nov 27 03:50:31 localhost podman[98042]: 2025-11-27 08:50:31.154879743 +0000 UTC m=+0.235691205 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, tcib_managed=true, io.openshift.expose-services=, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:50:31 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:50:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:50:33 localhost podman[98112]: 2025-11-27 08:50:33.9743713 +0000 UTC m=+0.074352315 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, config_id=tripleo_step4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vcs-type=git) Nov 27 03:50:34 localhost podman[98112]: 2025-11-27 08:50:34.357026607 +0000 UTC m=+0.457007632 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, vcs-type=git, tcib_managed=true, batch=17.1_20251118.1, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=) Nov 27 03:50:34 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:50:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:50:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:50:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:50:36 localhost systemd[1]: tmp-crun.GA1rE9.mount: Deactivated successfully. Nov 27 03:50:37 localhost podman[98137]: 2025-11-27 08:50:37.005106986 +0000 UTC m=+0.092626786 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, architecture=x86_64, release=1761123044, io.openshift.expose-services=, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, container_name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, vcs-type=git) Nov 27 03:50:37 localhost systemd[1]: tmp-crun.kvSvIR.mount: Deactivated successfully. Nov 27 03:50:37 localhost podman[98135]: 2025-11-27 08:50:37.054865061 +0000 UTC m=+0.145755551 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, container_name=ovn_metadata_agent, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:50:37 localhost podman[98137]: 2025-11-27 08:50:37.082745609 +0000 UTC m=+0.170265359 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, url=https://www.redhat.com, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute) Nov 27 03:50:37 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:50:37 localhost podman[98135]: 2025-11-27 08:50:37.093918819 +0000 UTC m=+0.184809279 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, batch=17.1_20251118.1, container_name=ovn_metadata_agent, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, build-date=2025-11-19T00:14:25Z, release=1761123044, managed_by=tripleo_ansible, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git) Nov 27 03:50:37 localhost podman[98135]: unhealthy Nov 27 03:50:37 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:50:37 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:50:37 localhost podman[98136]: 2025-11-27 08:50:37.099599282 +0000 UTC m=+0.188555801 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_controller, tcib_managed=true, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z) Nov 27 03:50:37 localhost podman[98136]: 2025-11-27 08:50:37.184056657 +0000 UTC m=+0.273013156 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible, config_id=tripleo_step4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, url=https://www.redhat.com, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:50:37 localhost podman[98136]: unhealthy Nov 27 03:50:37 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:50:37 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:50:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:50:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:50:44 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:50:44 localhost recover_tripleo_nova_virtqemud[98205]: 63639 Nov 27 03:50:44 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:50:44 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:50:45 localhost systemd[1]: tmp-crun.364jG0.mount: Deactivated successfully. Nov 27 03:50:45 localhost podman[98197]: 2025-11-27 08:50:45.011269514 +0000 UTC m=+0.101100924 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-iscsid-container, version=17.1.12, url=https://www.redhat.com, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, config_id=tripleo_step3, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, io.openshift.expose-services=) Nov 27 03:50:45 localhost podman[98197]: 2025-11-27 08:50:45.046370666 +0000 UTC m=+0.136202096 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, tcib_managed=true, io.openshift.expose-services=, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, distribution-scope=public, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:50:45 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:50:45 localhost podman[98196]: 2025-11-27 08:50:45.105363849 +0000 UTC m=+0.197932722 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., vcs-type=git, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, version=17.1.12, release=1761123044, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:50:45 localhost podman[98196]: 2025-11-27 08:50:45.119860598 +0000 UTC m=+0.212429521 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, name=rhosp17/openstack-collectd, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, vcs-type=git, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.buildah.version=1.41.4, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-collectd-container, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:50:45 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:50:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:50:56 localhost systemd[1]: tmp-crun.wyrZDy.mount: Deactivated successfully. Nov 27 03:50:57 localhost podman[98312]: 2025-11-27 08:50:56.999129292 +0000 UTC m=+0.089107122 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.buildah.version=1.41.4, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, vcs-type=git, version=17.1.12, url=https://www.redhat.com, architecture=x86_64, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:50:57 localhost podman[98312]: 2025-11-27 08:50:57.184901467 +0000 UTC m=+0.274879287 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, maintainer=OpenStack TripleO Team, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, batch=17.1_20251118.1, io.openshift.expose-services=, version=17.1.12, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible) Nov 27 03:50:57 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:51:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:51:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:51:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:51:02 localhost podman[98343]: 2025-11-27 08:51:01.998882926 +0000 UTC m=+0.086390619 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.4, version=17.1.12, tcib_managed=true, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, batch=17.1_20251118.1, architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, release=1761123044, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:51:02 localhost podman[98343]: 2025-11-27 08:51:02.034881732 +0000 UTC m=+0.122389425 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, architecture=x86_64, name=rhosp17/openstack-ceilometer-compute, container_name=ceilometer_agent_compute, vcs-type=git, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=) Nov 27 03:51:02 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:51:02 localhost podman[98341]: 2025-11-27 08:51:02.0497304 +0000 UTC m=+0.142121364 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, vcs-type=git, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, batch=17.1_20251118.1, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, container_name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true) Nov 27 03:51:02 localhost podman[98341]: 2025-11-27 08:51:02.10004425 +0000 UTC m=+0.192435184 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, version=17.1.12, maintainer=OpenStack TripleO Team, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, config_id=tripleo_step4, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true) Nov 27 03:51:02 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:51:02 localhost podman[98342]: 2025-11-27 08:51:02.107828519 +0000 UTC m=+0.197877670 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, container_name=logrotate_crond, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-cron, architecture=x86_64, build-date=2025-11-18T22:49:32Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:51:02 localhost podman[98342]: 2025-11-27 08:51:02.191942346 +0000 UTC m=+0.281991477 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, release=1761123044, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., url=https://www.redhat.com, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4) Nov 27 03:51:02 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:51:02 localhost systemd[1]: tmp-crun.joYjPB.mount: Deactivated successfully. Nov 27 03:51:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:51:04 localhost systemd[1]: tmp-crun.cdJ0L5.mount: Deactivated successfully. Nov 27 03:51:04 localhost podman[98412]: 2025-11-27 08:51:04.986766584 +0000 UTC m=+0.087336605 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., container_name=nova_migration_target, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, url=https://www.redhat.com, config_id=tripleo_step4, batch=17.1_20251118.1, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:51:05 localhost podman[98412]: 2025-11-27 08:51:05.289893946 +0000 UTC m=+0.390463967 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, url=https://www.redhat.com, release=1761123044, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, architecture=x86_64, container_name=nova_migration_target, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:51:05 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:51:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:51:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:51:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:51:08 localhost podman[98437]: 2025-11-27 08:51:08.005582079 +0000 UTC m=+0.093201371 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, config_id=tripleo_step5, io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:51:08 localhost podman[98435]: 2025-11-27 08:51:08.039401577 +0000 UTC m=+0.132891486 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, config_id=tripleo_step4, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_metadata_agent, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, architecture=x86_64, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 03:51:08 localhost podman[98435]: 2025-11-27 08:51:08.057848532 +0000 UTC m=+0.151338401 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, release=1761123044, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, tcib_managed=true) Nov 27 03:51:08 localhost systemd[1]: tmp-crun.p7iJve.mount: Deactivated successfully. Nov 27 03:51:08 localhost podman[98436]: 2025-11-27 08:51:08.099155421 +0000 UTC m=+0.190149433 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, version=17.1.12, distribution-scope=public, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, tcib_managed=true, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 03:51:08 localhost podman[98435]: unhealthy Nov 27 03:51:08 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:51:08 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:51:08 localhost podman[98436]: 2025-11-27 08:51:08.138240849 +0000 UTC m=+0.229234901 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, container_name=ovn_controller, release=1761123044, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12) Nov 27 03:51:08 localhost podman[98436]: unhealthy Nov 27 03:51:08 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:51:08 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:51:08 localhost podman[98437]: 2025-11-27 08:51:08.16548877 +0000 UTC m=+0.253108092 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, tcib_managed=true, version=17.1.12, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, config_id=tripleo_step5, vcs-type=git, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:51:08 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:51:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:51:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:51:15 localhost systemd[1]: tmp-crun.SizeyQ.mount: Deactivated successfully. Nov 27 03:51:15 localhost podman[98500]: 2025-11-27 08:51:15.996593691 +0000 UTC m=+0.091795004 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-type=git, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, container_name=collectd, distribution-scope=public, version=17.1.12, tcib_managed=true, managed_by=tripleo_ansible, config_id=tripleo_step3, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:51:16 localhost podman[98501]: 2025-11-27 08:51:16.044574578 +0000 UTC m=+0.134648763 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., version=17.1.12, container_name=iscsid, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, architecture=x86_64) Nov 27 03:51:16 localhost podman[98501]: 2025-11-27 08:51:16.056801536 +0000 UTC m=+0.146875721 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-iscsid, tcib_managed=true, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, vcs-type=git, com.redhat.component=openstack-iscsid-container, managed_by=tripleo_ansible, url=https://www.redhat.com, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:51:16 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:51:16 localhost podman[98500]: 2025-11-27 08:51:16.114348721 +0000 UTC m=+0.209550004 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_id=tripleo_step3, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.openshift.expose-services=, container_name=collectd, tcib_managed=true, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, distribution-scope=public, maintainer=OpenStack TripleO Team, release=1761123044, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, vcs-type=git) Nov 27 03:51:16 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:51:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:51:27 localhost systemd[1]: tmp-crun.6n4hNV.mount: Deactivated successfully. Nov 27 03:51:28 localhost podman[98541]: 2025-11-27 08:51:28.001477677 +0000 UTC m=+0.090873929 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, com.redhat.component=openstack-qdrouterd-container, release=1761123044, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, managed_by=tripleo_ansible, config_id=tripleo_step1, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, name=rhosp17/openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, tcib_managed=true, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com) Nov 27 03:51:28 localhost podman[98541]: 2025-11-27 08:51:28.217957795 +0000 UTC m=+0.307353997 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, config_id=tripleo_step1, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, distribution-scope=public, vendor=Red Hat, Inc.) Nov 27 03:51:28 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:51:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:51:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:51:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:51:32 localhost systemd[1]: tmp-crun.V64HIN.mount: Deactivated successfully. Nov 27 03:51:33 localhost podman[98572]: 2025-11-27 08:51:33.004431018 +0000 UTC m=+0.092880253 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, version=17.1.12, container_name=ceilometer_agent_compute, tcib_managed=true, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, architecture=x86_64, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, url=https://www.redhat.com, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible) Nov 27 03:51:33 localhost podman[98572]: 2025-11-27 08:51:33.037036063 +0000 UTC m=+0.125485298 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, tcib_managed=true, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, vcs-type=git) Nov 27 03:51:33 localhost podman[98571]: 2025-11-27 08:51:33.050638937 +0000 UTC m=+0.138913397 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., managed_by=tripleo_ansible, architecture=x86_64, version=17.1.12, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, maintainer=OpenStack TripleO Team, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, container_name=logrotate_crond, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=) Nov 27 03:51:33 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:51:33 localhost podman[98571]: 2025-11-27 08:51:33.087831846 +0000 UTC m=+0.176106296 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, container_name=logrotate_crond, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.buildah.version=1.41.4, config_id=tripleo_step4, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:51:33 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:51:33 localhost podman[98570]: 2025-11-27 08:51:33.102344495 +0000 UTC m=+0.192243599 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-type=git, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, build-date=2025-11-19T00:12:45Z, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, architecture=x86_64) Nov 27 03:51:33 localhost podman[98570]: 2025-11-27 08:51:33.135884275 +0000 UTC m=+0.225783439 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, architecture=x86_64, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, io.buildah.version=1.41.4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z) Nov 27 03:51:33 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:51:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:51:35 localhost systemd[1]: tmp-crun.Wx48nK.mount: Deactivated successfully. Nov 27 03:51:35 localhost podman[98645]: 2025-11-27 08:51:35.991861791 +0000 UTC m=+0.090113579 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, distribution-scope=public, vcs-type=git, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:51:36 localhost podman[98645]: 2025-11-27 08:51:36.357280436 +0000 UTC m=+0.455532234 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, container_name=nova_migration_target, url=https://www.redhat.com, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1) Nov 27 03:51:36 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:51:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:51:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:51:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:51:38 localhost podman[98668]: 2025-11-27 08:51:38.99550633 +0000 UTC m=+0.084772016 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, tcib_managed=true, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vcs-type=git, architecture=x86_64, build-date=2025-11-19T00:14:25Z, vendor=Red Hat, Inc.) Nov 27 03:51:39 localhost podman[98669]: 2025-11-27 08:51:39.05329276 +0000 UTC m=+0.139423492 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.openshift.expose-services=, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, batch=17.1_20251118.1, container_name=ovn_controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, architecture=x86_64, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc.) Nov 27 03:51:39 localhost podman[98670]: 2025-11-27 08:51:39.025140495 +0000 UTC m=+0.104136355 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, container_name=nova_compute, distribution-scope=public, version=17.1.12, managed_by=tripleo_ansible, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, tcib_managed=true, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step5, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:51:39 localhost podman[98668]: 2025-11-27 08:51:39.079950176 +0000 UTC m=+0.169215902 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, config_id=tripleo_step4, io.openshift.expose-services=, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, version=17.1.12, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, container_name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:51:39 localhost podman[98668]: unhealthy Nov 27 03:51:39 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:51:39 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:51:39 localhost podman[98669]: 2025-11-27 08:51:39.097783994 +0000 UTC m=+0.183914676 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, tcib_managed=true, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-type=git, config_id=tripleo_step4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 03:51:39 localhost podman[98669]: unhealthy Nov 27 03:51:39 localhost podman[98670]: 2025-11-27 08:51:39.111065641 +0000 UTC m=+0.190061531 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., version=17.1.12, tcib_managed=true, url=https://www.redhat.com, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, release=1761123044, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, batch=17.1_20251118.1, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team) Nov 27 03:51:39 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:51:39 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:51:39 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:51:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:51:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:51:46 localhost systemd[1]: tmp-crun.hpQHxg.mount: Deactivated successfully. Nov 27 03:51:47 localhost podman[98736]: 2025-11-27 08:51:47.050205301 +0000 UTC m=+0.136987847 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, distribution-scope=public, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, config_id=tripleo_step3, batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, container_name=iscsid) Nov 27 03:51:47 localhost podman[98736]: 2025-11-27 08:51:47.060804085 +0000 UTC m=+0.147586601 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, distribution-scope=public, io.buildah.version=1.41.4, build-date=2025-11-18T23:44:13Z, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-type=git, url=https://www.redhat.com, managed_by=tripleo_ansible, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, com.redhat.component=openstack-iscsid-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:51:47 localhost podman[98735]: 2025-11-27 08:51:47.021379488 +0000 UTC m=+0.110424285 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, tcib_managed=true, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, url=https://www.redhat.com, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}) Nov 27 03:51:47 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:51:47 localhost podman[98735]: 2025-11-27 08:51:47.105089333 +0000 UTC m=+0.194134090 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, url=https://www.redhat.com, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, config_id=tripleo_step3, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., distribution-scope=public, release=1761123044) Nov 27 03:51:47 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:51:49 localhost systemd[1]: tmp-crun.4m681G.mount: Deactivated successfully. Nov 27 03:51:49 localhost podman[98877]: 2025-11-27 08:51:49.795930159 +0000 UTC m=+0.093712285 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, release=553, RELEASE=main, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, ceph=True, name=rhceph, vcs-type=git, distribution-scope=public, io.buildah.version=1.33.12, GIT_BRANCH=main, architecture=x86_64, io.openshift.expose-services=, GIT_CLEAN=True) Nov 27 03:51:49 localhost podman[98877]: 2025-11-27 08:51:49.922005452 +0000 UTC m=+0.219787588 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, release=553, name=rhceph, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, RELEASE=main, io.buildah.version=1.33.12) Nov 27 03:51:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:51:59 localhost podman[99022]: 2025-11-27 08:51:59.000997846 +0000 UTC m=+0.090235642 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, tcib_managed=true, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-type=git, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, io.openshift.expose-services=) Nov 27 03:51:59 localhost podman[99022]: 2025-11-27 08:51:59.213097197 +0000 UTC m=+0.302335023 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:51:59 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:52:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:52:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:52:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:52:03 localhost systemd[1]: tmp-crun.7TlvpU.mount: Deactivated successfully. Nov 27 03:52:03 localhost podman[99053]: 2025-11-27 08:52:03.99622817 +0000 UTC m=+0.090586181 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, io.buildah.version=1.41.4, distribution-scope=public, vcs-type=git, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z, url=https://www.redhat.com, tcib_managed=true, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, config_id=tripleo_step4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron) Nov 27 03:52:04 localhost systemd[1]: tmp-crun.Hsk77o.mount: Deactivated successfully. Nov 27 03:52:04 localhost podman[99052]: 2025-11-27 08:52:04.040339774 +0000 UTC m=+0.136058622 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.expose-services=, url=https://www.redhat.com, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, tcib_managed=true, architecture=x86_64, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:52:04 localhost podman[99054]: 2025-11-27 08:52:04.09422753 +0000 UTC m=+0.185120698 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, distribution-scope=public, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.openshift.expose-services=, vcs-type=git, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, architecture=x86_64, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:52:04 localhost podman[99053]: 2025-11-27 08:52:04.112042467 +0000 UTC m=+0.206400438 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, architecture=x86_64, io.buildah.version=1.41.4, config_id=tripleo_step4, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., managed_by=tripleo_ansible) Nov 27 03:52:04 localhost podman[99052]: 2025-11-27 08:52:04.12220877 +0000 UTC m=+0.217927608 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:52:04 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:52:04 localhost podman[99054]: 2025-11-27 08:52:04.134793817 +0000 UTC m=+0.225687025 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, vcs-type=git, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, build-date=2025-11-19T00:11:48Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., batch=17.1_20251118.1, release=1761123044, distribution-scope=public, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute) Nov 27 03:52:04 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:52:04 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:52:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:52:06 localhost podman[99122]: 2025-11-27 08:52:06.971349354 +0000 UTC m=+0.071995853 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, container_name=nova_migration_target, config_id=tripleo_step4, distribution-scope=public, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:52:07 localhost podman[99122]: 2025-11-27 08:52:07.364998426 +0000 UTC m=+0.465644855 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, build-date=2025-11-19T00:36:58Z, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_id=tripleo_step4, managed_by=tripleo_ansible, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public) Nov 27 03:52:07 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:52:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:52:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:52:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:52:09 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:52:09 localhost recover_tripleo_nova_virtqemud[99157]: 63639 Nov 27 03:52:09 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:52:09 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:52:09 localhost podman[99145]: 2025-11-27 08:52:09.992441881 +0000 UTC m=+0.090185161 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, io.buildah.version=1.41.4, distribution-scope=public, container_name=ovn_metadata_agent, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, architecture=x86_64, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, config_id=tripleo_step4, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 03:52:10 localhost podman[99145]: 2025-11-27 08:52:10.006840168 +0000 UTC m=+0.104583438 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible) Nov 27 03:52:10 localhost podman[99145]: unhealthy Nov 27 03:52:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:52:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:52:10 localhost podman[99147]: 2025-11-27 08:52:10.051194517 +0000 UTC m=+0.137902031 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, config_id=tripleo_step5, distribution-scope=public, build-date=2025-11-19T00:36:58Z, tcib_managed=true, release=1761123044, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, container_name=nova_compute, name=rhosp17/openstack-nova-compute, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:52:10 localhost podman[99146]: 2025-11-27 08:52:10.096064211 +0000 UTC m=+0.186978207 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4) Nov 27 03:52:10 localhost podman[99146]: 2025-11-27 08:52:10.11389778 +0000 UTC m=+0.204811836 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, io.buildah.version=1.41.4, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, url=https://www.redhat.com, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-type=git, distribution-scope=public, maintainer=OpenStack TripleO Team, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true) Nov 27 03:52:10 localhost podman[99147]: 2025-11-27 08:52:10.114369882 +0000 UTC m=+0.201077426 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, distribution-scope=public, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git) Nov 27 03:52:10 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:52:10 localhost podman[99146]: unhealthy Nov 27 03:52:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:52:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:52:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:52:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:52:17 localhost podman[99213]: 2025-11-27 08:52:17.991765167 +0000 UTC m=+0.086477261 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-18T22:51:28Z, tcib_managed=true, vcs-type=git, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=collectd, config_id=tripleo_step3, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible) Nov 27 03:52:18 localhost podman[99213]: 2025-11-27 08:52:18.00082316 +0000 UTC m=+0.095535284 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, tcib_managed=true, config_id=tripleo_step3, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, build-date=2025-11-18T22:51:28Z, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:52:18 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:52:18 localhost podman[99214]: 2025-11-27 08:52:18.041511702 +0000 UTC m=+0.133748670 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, version=17.1.12, architecture=x86_64, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, vcs-type=git, container_name=iscsid, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, batch=17.1_20251118.1, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:52:18 localhost podman[99214]: 2025-11-27 08:52:18.053828252 +0000 UTC m=+0.146065200 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, architecture=x86_64, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, io.buildah.version=1.41.4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, container_name=iscsid, url=https://www.redhat.com, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-type=git, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:52:18 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:52:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:52:29 localhost systemd[1]: tmp-crun.fs2TrP.mount: Deactivated successfully. Nov 27 03:52:30 localhost podman[99251]: 2025-11-27 08:52:30.002213842 +0000 UTC m=+0.093758397 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, version=17.1.12, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, container_name=metrics_qdr, distribution-scope=public, architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 03:52:30 localhost podman[99251]: 2025-11-27 08:52:30.194629114 +0000 UTC m=+0.286173709 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_id=tripleo_step1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible) Nov 27 03:52:30 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:52:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:52:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:52:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:52:34 localhost podman[99281]: 2025-11-27 08:52:34.989252535 +0000 UTC m=+0.084815596 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:52:35 localhost systemd[1]: tmp-crun.RcNplt.mount: Deactivated successfully. Nov 27 03:52:35 localhost podman[99282]: 2025-11-27 08:52:35.056274863 +0000 UTC m=+0.146424039 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, distribution-scope=public, container_name=logrotate_crond, config_id=tripleo_step4, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-type=git, architecture=x86_64, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:52:35 localhost podman[99282]: 2025-11-27 08:52:35.063186059 +0000 UTC m=+0.153335235 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, name=rhosp17/openstack-cron, url=https://www.redhat.com, distribution-scope=public, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, version=17.1.12, container_name=logrotate_crond, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z) Nov 27 03:52:35 localhost podman[99283]: 2025-11-27 08:52:35.019734553 +0000 UTC m=+0.106943910 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, io.openshift.expose-services=, version=17.1.12, io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, vendor=Red Hat, Inc., managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z) Nov 27 03:52:35 localhost podman[99281]: 2025-11-27 08:52:35.074224425 +0000 UTC m=+0.169787496 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., batch=17.1_20251118.1, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 03:52:35 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:52:35 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:52:35 localhost podman[99283]: 2025-11-27 08:52:35.10197906 +0000 UTC m=+0.189188377 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, distribution-scope=public, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, release=1761123044, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, version=17.1.12) Nov 27 03:52:35 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:52:35 localhost systemd[1]: tmp-crun.AiYNxj.mount: Deactivated successfully. Nov 27 03:52:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:52:37 localhost systemd[1]: tmp-crun.XlzCfr.mount: Deactivated successfully. Nov 27 03:52:37 localhost podman[99354]: 2025-11-27 08:52:37.99684813 +0000 UTC m=+0.096333965 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, managed_by=tripleo_ansible, version=17.1.12, build-date=2025-11-19T00:36:58Z, tcib_managed=true, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:52:38 localhost podman[99354]: 2025-11-27 08:52:38.371934934 +0000 UTC m=+0.471420799 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-type=git, tcib_managed=true, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vendor=Red Hat, Inc., container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, release=1761123044, name=rhosp17/openstack-nova-compute, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:52:38 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:52:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:52:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:52:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:52:40 localhost systemd[1]: tmp-crun.nLxxj5.mount: Deactivated successfully. Nov 27 03:52:40 localhost podman[99379]: 2025-11-27 08:52:40.995575866 +0000 UTC m=+0.089563123 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, vendor=Red Hat, Inc., managed_by=tripleo_ansible, vcs-type=git, build-date=2025-11-19T00:14:25Z) Nov 27 03:52:41 localhost podman[99379]: 2025-11-27 08:52:41.041695874 +0000 UTC m=+0.135683151 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-type=git) Nov 27 03:52:41 localhost podman[99379]: unhealthy Nov 27 03:52:41 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:52:41 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:52:41 localhost podman[99381]: 2025-11-27 08:52:41.041398886 +0000 UTC m=+0.130405710 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, version=17.1.12, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., config_id=tripleo_step5, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:52:41 localhost podman[99381]: 2025-11-27 08:52:41.124922617 +0000 UTC m=+0.213929411 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, container_name=nova_compute, version=17.1.12, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, distribution-scope=public, tcib_managed=true) Nov 27 03:52:41 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:52:41 localhost podman[99380]: 2025-11-27 08:52:41.130097316 +0000 UTC m=+0.221388712 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, distribution-scope=public, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_id=tripleo_step4, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:52:41 localhost podman[99380]: 2025-11-27 08:52:41.213964175 +0000 UTC m=+0.305255571 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1, config_id=tripleo_step4, vendor=Red Hat, Inc., vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:52:41 localhost podman[99380]: unhealthy Nov 27 03:52:41 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:52:41 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:52:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:52:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:52:48 localhost systemd[1]: tmp-crun.CEx5YK.mount: Deactivated successfully. Nov 27 03:52:48 localhost podman[99443]: 2025-11-27 08:52:48.992861505 +0000 UTC m=+0.091841015 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, container_name=collectd, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:52:49 localhost podman[99443]: 2025-11-27 08:52:49.005896685 +0000 UTC m=+0.104876175 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, architecture=x86_64, vcs-type=git, distribution-scope=public, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, version=17.1.12, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:52:49 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:52:49 localhost podman[99444]: 2025-11-27 08:52:49.088846631 +0000 UTC m=+0.185129239 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, architecture=x86_64, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., url=https://www.redhat.com, config_id=tripleo_step3, tcib_managed=true, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, managed_by=tripleo_ansible, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:52:49 localhost podman[99444]: 2025-11-27 08:52:49.101891911 +0000 UTC m=+0.198174569 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, distribution-scope=public, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, release=1761123044, io.openshift.expose-services=, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, architecture=x86_64, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 03:52:49 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:53:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:53:01 localhost podman[99559]: 2025-11-27 08:53:01.003480655 +0000 UTC m=+0.094986890 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, name=rhosp17/openstack-qdrouterd, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_step1, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, vendor=Red Hat, Inc., release=1761123044, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:53:01 localhost podman[99559]: 2025-11-27 08:53:01.195880036 +0000 UTC m=+0.287386221 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, version=17.1.12, vendor=Red Hat, Inc., managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, architecture=x86_64, vcs-type=git, io.buildah.version=1.41.4) Nov 27 03:53:01 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:53:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:53:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:53:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:53:05 localhost podman[99587]: 2025-11-27 08:53:05.978008062 +0000 UTC m=+0.068924999 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, version=17.1.12, managed_by=tripleo_ansible, config_id=tripleo_step4, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, vcs-type=git, batch=17.1_20251118.1, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, release=1761123044, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:53:06 localhost podman[99587]: 2025-11-27 08:53:06.015720785 +0000 UTC m=+0.106637692 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, managed_by=tripleo_ansible, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, tcib_managed=true, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, architecture=x86_64, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.component=openstack-cron-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-cron, distribution-scope=public, io.openshift.expose-services=) Nov 27 03:53:06 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:53:06 localhost podman[99586]: 2025-11-27 08:53:06.089952386 +0000 UTC m=+0.183355130 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, vcs-type=git, config_id=tripleo_step4, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, batch=17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:53:06 localhost podman[99586]: 2025-11-27 08:53:06.125700335 +0000 UTC m=+0.219103079 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, release=1761123044, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-ipmi-container, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, version=17.1.12, vcs-type=git, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:53:06 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:53:06 localhost podman[99588]: 2025-11-27 08:53:06.145758443 +0000 UTC m=+0.236934147 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:53:06 localhost podman[99588]: 2025-11-27 08:53:06.174477554 +0000 UTC m=+0.265653228 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, name=rhosp17/openstack-ceilometer-compute, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_compute, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., version=17.1.12, build-date=2025-11-19T00:11:48Z, io.buildah.version=1.41.4, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:53:06 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:53:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:53:08 localhost systemd[1]: tmp-crun.AOLIIm.mount: Deactivated successfully. Nov 27 03:53:09 localhost podman[99659]: 2025-11-27 08:53:09.001737951 +0000 UTC m=+0.096365147 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:53:09 localhost podman[99659]: 2025-11-27 08:53:09.388433496 +0000 UTC m=+0.483060662 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, container_name=nova_migration_target, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:53:09 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:53:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:53:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:53:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:53:11 localhost systemd[1]: tmp-crun.z6asFn.mount: Deactivated successfully. Nov 27 03:53:11 localhost podman[99683]: 2025-11-27 08:53:11.998643779 +0000 UTC m=+0.095507024 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true) Nov 27 03:53:12 localhost podman[99685]: 2025-11-27 08:53:12.042035603 +0000 UTC m=+0.134253574 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.openshift.expose-services=, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, batch=17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, vcs-type=git, tcib_managed=true, config_id=tripleo_step5, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:53:12 localhost podman[99683]: 2025-11-27 08:53:12.043446701 +0000 UTC m=+0.140309946 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, version=17.1.12, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:53:12 localhost podman[99683]: unhealthy Nov 27 03:53:12 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:53:12 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:53:12 localhost podman[99684]: 2025-11-27 08:53:12.109269176 +0000 UTC m=+0.202342499 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, version=17.1.12, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, tcib_managed=true, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, io.openshift.expose-services=, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:53:12 localhost podman[99685]: 2025-11-27 08:53:12.124947238 +0000 UTC m=+0.217165219 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, version=17.1.12, build-date=2025-11-19T00:36:58Z, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, container_name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, config_id=tripleo_step5, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, distribution-scope=public, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:53:12 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:53:12 localhost podman[99684]: 2025-11-27 08:53:12.152841176 +0000 UTC m=+0.245914529 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, architecture=x86_64, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, version=17.1.12, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, io.buildah.version=1.41.4, release=1761123044, batch=17.1_20251118.1, tcib_managed=true, url=https://www.redhat.com) Nov 27 03:53:12 localhost podman[99684]: unhealthy Nov 27 03:53:12 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:53:12 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:53:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:53:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:53:19 localhost systemd[1]: tmp-crun.8zA21r.mount: Deactivated successfully. Nov 27 03:53:20 localhost podman[99749]: 2025-11-27 08:53:20.006682567 +0000 UTC m=+0.097751133 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, com.redhat.component=openstack-collectd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, version=17.1.12, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:53:20 localhost podman[99750]: 2025-11-27 08:53:20.050660758 +0000 UTC m=+0.137080159 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, container_name=iscsid, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, config_id=tripleo_step3, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:53:20 localhost podman[99750]: 2025-11-27 08:53:20.064880249 +0000 UTC m=+0.151299680 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, config_id=tripleo_step3, release=1761123044, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., container_name=iscsid, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:53:20 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:53:20 localhost podman[99749]: 2025-11-27 08:53:20.117495021 +0000 UTC m=+0.208563537 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, batch=17.1_20251118.1, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, container_name=collectd, io.buildah.version=1.41.4, version=17.1.12, io.openshift.expose-services=, release=1761123044, build-date=2025-11-18T22:51:28Z, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:53:20 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:53:20 localhost systemd[1]: tmp-crun.RAuXkI.mount: Deactivated successfully. Nov 27 03:53:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:53:31 localhost systemd[1]: tmp-crun.gVFlvJ.mount: Deactivated successfully. Nov 27 03:53:31 localhost podman[99788]: 2025-11-27 08:53:31.994400162 +0000 UTC m=+0.093597452 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, batch=17.1_20251118.1, config_id=tripleo_step1, managed_by=tripleo_ansible, distribution-scope=public, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.expose-services=) Nov 27 03:53:32 localhost podman[99788]: 2025-11-27 08:53:32.182005215 +0000 UTC m=+0.281202515 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, tcib_managed=true, io.openshift.expose-services=, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd) Nov 27 03:53:32 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:53:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:53:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:53:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:53:36 localhost podman[99818]: 2025-11-27 08:53:36.987580832 +0000 UTC m=+0.080928052 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, batch=17.1_20251118.1, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, version=17.1.12, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 03:53:37 localhost podman[99818]: 2025-11-27 08:53:37.035109827 +0000 UTC m=+0.128457057 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, config_id=tripleo_step4, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.buildah.version=1.41.4, vcs-type=git, managed_by=tripleo_ansible, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:53:37 localhost systemd[1]: tmp-crun.NUYA68.mount: Deactivated successfully. Nov 27 03:53:37 localhost podman[99820]: 2025-11-27 08:53:37.058829033 +0000 UTC m=+0.143083260 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.expose-services=, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, architecture=x86_64, version=17.1.12, batch=17.1_20251118.1, tcib_managed=true, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:53:37 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:53:37 localhost podman[99820]: 2025-11-27 08:53:37.082760456 +0000 UTC m=+0.167014683 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, vcs-type=git, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, release=1761123044, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:53:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:53:37 localhost podman[99819]: 2025-11-27 08:53:37.097350637 +0000 UTC m=+0.183905706 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, vcs-type=git, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=logrotate_crond, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, distribution-scope=public, com.redhat.component=openstack-cron-container, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, architecture=x86_64, io.buildah.version=1.41.4, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 03:53:37 localhost podman[99819]: 2025-11-27 08:53:37.135627155 +0000 UTC m=+0.222182234 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, vcs-type=git, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, architecture=x86_64, tcib_managed=true, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, io.buildah.version=1.41.4, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:53:37 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:53:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:53:39 localhost systemd[1]: tmp-crun.bAVBca.mount: Deactivated successfully. Nov 27 03:53:40 localhost podman[99891]: 2025-11-27 08:53:40.001866717 +0000 UTC m=+0.098678829 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, container_name=nova_migration_target, batch=17.1_20251118.1, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:53:40 localhost podman[99891]: 2025-11-27 08:53:40.369587862 +0000 UTC m=+0.466399974 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, version=17.1.12, distribution-scope=public, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_id=tripleo_step4, url=https://www.redhat.com, release=1761123044, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:53:40 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:53:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:53:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:53:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:53:42 localhost podman[99916]: 2025-11-27 08:53:42.9814444 +0000 UTC m=+0.074532100 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_controller, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, architecture=x86_64, tcib_managed=true) Nov 27 03:53:43 localhost podman[99916]: 2025-11-27 08:53:43.022372448 +0000 UTC m=+0.115460098 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, release=1761123044, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_id=tripleo_step4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, distribution-scope=public, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ovn-controller) Nov 27 03:53:43 localhost systemd[1]: tmp-crun.AJZ63u.mount: Deactivated successfully. Nov 27 03:53:43 localhost podman[99916]: unhealthy Nov 27 03:53:43 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:53:43 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:53:43 localhost podman[99917]: 2025-11-27 08:53:43.081792092 +0000 UTC m=+0.170632069 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, tcib_managed=true, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12) Nov 27 03:53:43 localhost podman[99915]: 2025-11-27 08:53:43.03773204 +0000 UTC m=+0.133075921 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, tcib_managed=true, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12) Nov 27 03:53:43 localhost podman[99917]: 2025-11-27 08:53:43.109870336 +0000 UTC m=+0.198710383 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_id=tripleo_step5, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.openshift.expose-services=, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, version=17.1.12, io.buildah.version=1.41.4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, container_name=nova_compute, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible) Nov 27 03:53:43 localhost podman[99915]: 2025-11-27 08:53:43.120992074 +0000 UTC m=+0.216335955 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, vcs-type=git, version=17.1.12, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, url=https://www.redhat.com) Nov 27 03:53:43 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:53:43 localhost podman[99915]: unhealthy Nov 27 03:53:43 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:53:43 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:53:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:53:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:53:50 localhost podman[99979]: 2025-11-27 08:53:50.986692203 +0000 UTC m=+0.082321549 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_id=tripleo_step3, release=1761123044, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-type=git, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., tcib_managed=true) Nov 27 03:53:50 localhost podman[99979]: 2025-11-27 08:53:50.99476353 +0000 UTC m=+0.090392886 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, release=1761123044, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd) Nov 27 03:53:51 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:53:51 localhost podman[99980]: 2025-11-27 08:53:51.039383787 +0000 UTC m=+0.132243659 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, container_name=iscsid, name=rhosp17/openstack-iscsid, version=17.1.12, build-date=2025-11-18T23:44:13Z, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, release=1761123044, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:53:51 localhost podman[99980]: 2025-11-27 08:53:51.052049416 +0000 UTC m=+0.144909288 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vendor=Red Hat, Inc., url=https://www.redhat.com, tcib_managed=true, com.redhat.component=openstack-iscsid-container, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=iscsid, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-type=git, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4) Nov 27 03:53:51 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:54:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:54:02 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:54:02 localhost recover_tripleo_nova_virtqemud[100098]: 63639 Nov 27 03:54:02 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:54:02 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:54:02 localhost podman[100096]: 2025-11-27 08:54:02.956251141 +0000 UTC m=+0.057714309 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, distribution-scope=public, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, io.openshift.expose-services=, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_id=tripleo_step1, tcib_managed=true, version=17.1.12) Nov 27 03:54:03 localhost podman[100096]: 2025-11-27 08:54:03.137716259 +0000 UTC m=+0.239179487 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, tcib_managed=true, version=17.1.12, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, io.buildah.version=1.41.4, config_id=tripleo_step1, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:54:03 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:54:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:54:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:54:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:54:07 localhost podman[100128]: 2025-11-27 08:54:07.992695791 +0000 UTC m=+0.087495719 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, architecture=x86_64, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:12:45Z, name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, tcib_managed=true, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, version=17.1.12) Nov 27 03:54:08 localhost podman[100128]: 2025-11-27 08:54:08.058961459 +0000 UTC m=+0.153761377 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, io.openshift.expose-services=, url=https://www.redhat.com, managed_by=tripleo_ansible, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, architecture=x86_64, distribution-scope=public, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:54:08 localhost podman[100129]: 2025-11-27 08:54:08.069636585 +0000 UTC m=+0.161221337 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, architecture=x86_64, config_id=tripleo_step4, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, tcib_managed=true, name=rhosp17/openstack-cron, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public) Nov 27 03:54:08 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:54:08 localhost podman[100129]: 2025-11-27 08:54:08.079765657 +0000 UTC m=+0.171350449 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, io.buildah.version=1.41.4, container_name=logrotate_crond, release=1761123044, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, distribution-scope=public, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:54:08 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:54:08 localhost podman[100130]: 2025-11-27 08:54:08.154309497 +0000 UTC m=+0.239930648 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, version=17.1.12, container_name=ceilometer_agent_compute, url=https://www.redhat.com, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, tcib_managed=true) Nov 27 03:54:08 localhost podman[100130]: 2025-11-27 08:54:08.186341107 +0000 UTC m=+0.271962258 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, name=rhosp17/openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, config_id=tripleo_step4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, tcib_managed=true, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, distribution-scope=public, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, build-date=2025-11-19T00:11:48Z, architecture=x86_64) Nov 27 03:54:08 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:54:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:54:10 localhost systemd[1]: tmp-crun.mZhNVd.mount: Deactivated successfully. Nov 27 03:54:11 localhost podman[100200]: 2025-11-27 08:54:11.000449039 +0000 UTC m=+0.095906944 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, version=17.1.12, container_name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, release=1761123044, vendor=Red Hat, Inc., batch=17.1_20251118.1, managed_by=tripleo_ansible, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, vcs-type=git, config_id=tripleo_step4, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:54:11 localhost podman[100200]: 2025-11-27 08:54:11.402883807 +0000 UTC m=+0.498341702 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.expose-services=, architecture=x86_64, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, container_name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, tcib_managed=true, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:54:11 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:54:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:54:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:54:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:54:13 localhost systemd[1]: tmp-crun.uFXRDX.mount: Deactivated successfully. Nov 27 03:54:14 localhost podman[100224]: 2025-11-27 08:54:14.000933774 +0000 UTC m=+0.088676240 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_id=tripleo_step4, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, architecture=x86_64, name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, release=1761123044) Nov 27 03:54:14 localhost systemd[1]: tmp-crun.94DFIt.mount: Deactivated successfully. Nov 27 03:54:14 localhost podman[100223]: 2025-11-27 08:54:14.047245777 +0000 UTC m=+0.137905411 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, tcib_managed=true, managed_by=tripleo_ansible, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, batch=17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12) Nov 27 03:54:14 localhost podman[100224]: 2025-11-27 08:54:14.050892314 +0000 UTC m=+0.138634740 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, container_name=ovn_controller, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, architecture=x86_64, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, io.buildah.version=1.41.4) Nov 27 03:54:14 localhost podman[100224]: unhealthy Nov 27 03:54:14 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:54:14 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:54:14 localhost podman[100225]: 2025-11-27 08:54:14.100868045 +0000 UTC m=+0.185978230 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, release=1761123044) Nov 27 03:54:14 localhost podman[100223]: 2025-11-27 08:54:14.119719291 +0000 UTC m=+0.210378915 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, container_name=ovn_metadata_agent, release=1761123044, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 03:54:14 localhost podman[100223]: unhealthy Nov 27 03:54:14 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:54:14 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:54:14 localhost podman[100225]: 2025-11-27 08:54:14.155869151 +0000 UTC m=+0.240979336 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, architecture=x86_64, vendor=Red Hat, Inc., io.openshift.expose-services=, io.buildah.version=1.41.4, url=https://www.redhat.com, config_id=tripleo_step5, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, release=1761123044, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:54:14 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:54:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:54:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:54:21 localhost systemd[1]: tmp-crun.GvfQ89.mount: Deactivated successfully. Nov 27 03:54:22 localhost podman[100287]: 2025-11-27 08:54:22.004082493 +0000 UTC m=+0.100866028 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, name=rhosp17/openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, release=1761123044, architecture=x86_64, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, container_name=iscsid, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:54:22 localhost podman[100287]: 2025-11-27 08:54:22.043938382 +0000 UTC m=+0.140721907 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., config_id=tripleo_step3, name=rhosp17/openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, distribution-scope=public, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, tcib_managed=true, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=) Nov 27 03:54:22 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:54:22 localhost podman[100286]: 2025-11-27 08:54:22.092918426 +0000 UTC m=+0.189592468 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vcs-type=git, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, version=17.1.12, url=https://www.redhat.com) Nov 27 03:54:22 localhost podman[100286]: 2025-11-27 08:54:22.126055706 +0000 UTC m=+0.222729758 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, container_name=collectd, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, name=rhosp17/openstack-collectd, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc.) Nov 27 03:54:22 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:54:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:54:33 localhost systemd[1]: tmp-crun.Uxz9If.mount: Deactivated successfully. Nov 27 03:54:33 localhost podman[100325]: 2025-11-27 08:54:33.988659487 +0000 UTC m=+0.086527851 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, config_id=tripleo_step1, url=https://www.redhat.com, vendor=Red Hat, Inc., io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:46Z, version=17.1.12, distribution-scope=public, tcib_managed=true, managed_by=tripleo_ansible, architecture=x86_64, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:54:34 localhost podman[100325]: 2025-11-27 08:54:34.207997545 +0000 UTC m=+0.305865899 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, architecture=x86_64, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, io.openshift.expose-services=, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, config_id=tripleo_step1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git) Nov 27 03:54:34 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:54:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:54:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:54:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:54:39 localhost systemd[1]: tmp-crun.5uogNI.mount: Deactivated successfully. Nov 27 03:54:39 localhost podman[100355]: 2025-11-27 08:54:39.050479655 +0000 UTC m=+0.134033271 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, build-date=2025-11-18T22:49:32Z, name=rhosp17/openstack-cron, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, distribution-scope=public, batch=17.1_20251118.1, container_name=logrotate_crond, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron) Nov 27 03:54:39 localhost podman[100355]: 2025-11-27 08:54:39.058664766 +0000 UTC m=+0.142218352 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, tcib_managed=true, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, url=https://www.redhat.com, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, io.buildah.version=1.41.4) Nov 27 03:54:39 localhost podman[100356]: 2025-11-27 08:54:39.014716772 +0000 UTC m=+0.095646327 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, name=rhosp17/openstack-ceilometer-compute, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, vcs-type=git, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, tcib_managed=true, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:54:39 localhost podman[100356]: 2025-11-27 08:54:39.099067553 +0000 UTC m=+0.179997058 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, config_id=tripleo_step4, vcs-type=git, io.buildah.version=1.41.4, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1) Nov 27 03:54:39 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:54:39 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:54:39 localhost podman[100354]: 2025-11-27 08:54:39.20027222 +0000 UTC m=+0.290063554 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, distribution-scope=public, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., tcib_managed=true, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:54:39 localhost podman[100354]: 2025-11-27 08:54:39.227165834 +0000 UTC m=+0.316957198 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, architecture=x86_64, distribution-scope=public, release=1761123044, version=17.1.12, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, url=https://www.redhat.com, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:54:39 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:54:39 localhost systemd[1]: tmp-crun.VmFSQE.mount: Deactivated successfully. Nov 27 03:54:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:54:42 localhost podman[100427]: 2025-11-27 08:54:42.00169121 +0000 UTC m=+0.088569297 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, vcs-type=git, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.openshift.expose-services=, tcib_managed=true, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044) Nov 27 03:54:42 localhost podman[100427]: 2025-11-27 08:54:42.350093643 +0000 UTC m=+0.436971780 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com) Nov 27 03:54:42 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:54:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:54:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:54:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:54:44 localhost podman[100451]: 2025-11-27 08:54:44.983756714 +0000 UTC m=+0.070157561 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, distribution-scope=public, architecture=x86_64, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step5, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:54:45 localhost podman[100449]: 2025-11-27 08:54:45.05526687 +0000 UTC m=+0.145164961 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, tcib_managed=true, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-type=git, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.openshift.expose-services=, vendor=Red Hat, Inc.) Nov 27 03:54:45 localhost podman[100449]: 2025-11-27 08:54:45.10389836 +0000 UTC m=+0.193796411 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, architecture=x86_64, tcib_managed=true, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git) Nov 27 03:54:45 localhost podman[100449]: unhealthy Nov 27 03:54:45 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:54:45 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:54:45 localhost podman[100451]: 2025-11-27 08:54:45.119912921 +0000 UTC m=+0.206313758 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, release=1761123044, vcs-type=git, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, config_id=tripleo_step5) Nov 27 03:54:45 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:54:45 localhost podman[100450]: 2025-11-27 08:54:45.101985958 +0000 UTC m=+0.190084701 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., batch=17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, tcib_managed=true, architecture=x86_64, managed_by=tripleo_ansible, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, config_id=tripleo_step4, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, vcs-type=git) Nov 27 03:54:45 localhost podman[100450]: 2025-11-27 08:54:45.186421322 +0000 UTC m=+0.274520065 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, distribution-scope=public, tcib_managed=true, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, io.openshift.expose-services=, release=1761123044, version=17.1.12, vcs-type=git) Nov 27 03:54:45 localhost podman[100450]: unhealthy Nov 27 03:54:45 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:54:45 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:54:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:54:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:54:52 localhost systemd[1]: tmp-crun.drG3x1.mount: Deactivated successfully. Nov 27 03:54:52 localhost podman[100516]: 2025-11-27 08:54:52.998653686 +0000 UTC m=+0.086115541 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, vcs-type=git, config_id=tripleo_step3, version=17.1.12, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-iscsid-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vendor=Red Hat, Inc., io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, name=rhosp17/openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:54:53 localhost podman[100516]: 2025-11-27 08:54:53.030837523 +0000 UTC m=+0.118299308 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, container_name=iscsid, tcib_managed=true, batch=17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, architecture=x86_64) Nov 27 03:54:53 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:54:53 localhost podman[100515]: 2025-11-27 08:54:53.039913767 +0000 UTC m=+0.129272502 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vendor=Red Hat, Inc., batch=17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-collectd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, architecture=x86_64, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:54:53 localhost podman[100515]: 2025-11-27 08:54:53.123985991 +0000 UTC m=+0.213344706 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, tcib_managed=true, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, url=https://www.redhat.com, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., container_name=collectd, release=1761123044) Nov 27 03:54:53 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:55:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:55:04 localhost podman[100631]: 2025-11-27 08:55:04.995395179 +0000 UTC m=+0.088018882 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, architecture=x86_64, release=1761123044, container_name=metrics_qdr, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, distribution-scope=public, tcib_managed=true, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:55:05 localhost podman[100631]: 2025-11-27 08:55:05.188934261 +0000 UTC m=+0.281557984 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., version=17.1.12, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, distribution-scope=public, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.openshift.expose-services=, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.component=openstack-qdrouterd-container) Nov 27 03:55:05 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:55:08 localhost sshd[100661]: main: sshd: ssh-rsa algorithm is disabled Nov 27 03:55:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:55:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:55:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:55:10 localhost systemd[1]: tmp-crun.i5oxfe.mount: Deactivated successfully. Nov 27 03:55:10 localhost podman[100665]: 2025-11-27 08:55:10.040928038 +0000 UTC m=+0.127887936 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vcs-type=git, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, tcib_managed=true, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc.) Nov 27 03:55:10 localhost podman[100665]: 2025-11-27 08:55:10.069119577 +0000 UTC m=+0.156079435 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-compute-container, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, vcs-type=git) Nov 27 03:55:10 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:55:10 localhost podman[100664]: 2025-11-27 08:55:10.115650439 +0000 UTC m=+0.204943140 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, architecture=x86_64, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, build-date=2025-11-18T22:49:32Z, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-cron) Nov 27 03:55:10 localhost podman[100664]: 2025-11-27 08:55:10.131792004 +0000 UTC m=+0.221084705 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, architecture=x86_64, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, vcs-type=git) Nov 27 03:55:10 localhost podman[100663]: 2025-11-27 08:55:09.994721563 +0000 UTC m=+0.089644506 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, tcib_managed=true, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4) Nov 27 03:55:10 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:55:10 localhost podman[100663]: 2025-11-27 08:55:10.174936456 +0000 UTC m=+0.269859439 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vcs-type=git, vendor=Red Hat, Inc., io.openshift.expose-services=, release=1761123044, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:55:10 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:55:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:55:12 localhost systemd[1]: tmp-crun.XemHUI.mount: Deactivated successfully. Nov 27 03:55:12 localhost podman[100735]: 2025-11-27 08:55:12.979754398 +0000 UTC m=+0.079575165 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, managed_by=tripleo_ansible, vcs-type=git, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, distribution-scope=public, name=rhosp17/openstack-nova-compute, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, vendor=Red Hat, Inc.) Nov 27 03:55:13 localhost podman[100735]: 2025-11-27 08:55:13.364794858 +0000 UTC m=+0.464615665 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, release=1761123044, build-date=2025-11-19T00:36:58Z, vcs-type=git, vendor=Red Hat, Inc., container_name=nova_migration_target, io.openshift.expose-services=, tcib_managed=true, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, version=17.1.12, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, config_id=tripleo_step4) Nov 27 03:55:13 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:55:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:55:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:55:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:55:15 localhost podman[100759]: 2025-11-27 08:55:15.997391331 +0000 UTC m=+0.090398265 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, architecture=x86_64, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:55:16 localhost podman[100759]: 2025-11-27 08:55:16.010734121 +0000 UTC m=+0.103741005 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.buildah.version=1.41.4, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, container_name=ovn_controller, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, release=1761123044) Nov 27 03:55:16 localhost podman[100759]: unhealthy Nov 27 03:55:16 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:55:16 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:55:16 localhost podman[100760]: 2025-11-27 08:55:16.059645037 +0000 UTC m=+0.145234932 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, url=https://www.redhat.com, container_name=nova_compute, release=1761123044, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:55:16 localhost podman[100758]: 2025-11-27 08:55:16.092843972 +0000 UTC m=+0.185364873 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., release=1761123044, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, managed_by=tripleo_ansible, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.buildah.version=1.41.4, container_name=ovn_metadata_agent, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:55:16 localhost podman[100758]: 2025-11-27 08:55:16.110709703 +0000 UTC m=+0.203230614 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, distribution-scope=public, url=https://www.redhat.com, version=17.1.12, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:55:16 localhost podman[100760]: 2025-11-27 08:55:16.111031071 +0000 UTC m=+0.196620936 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, architecture=x86_64, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vcs-type=git, release=1761123044, container_name=nova_compute, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, config_id=tripleo_step5) Nov 27 03:55:16 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:55:16 localhost podman[100758]: unhealthy Nov 27 03:55:16 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:55:16 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:55:16 localhost systemd[1]: tmp-crun.IZuPFA.mount: Deactivated successfully. Nov 27 03:55:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:55:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:55:23 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:55:23 localhost recover_tripleo_nova_virtqemud[100829]: 63639 Nov 27 03:55:23 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:55:23 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:55:23 localhost podman[100821]: 2025-11-27 08:55:23.994436112 +0000 UTC m=+0.088100234 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, container_name=collectd, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., release=1761123044, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.buildah.version=1.41.4, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public) Nov 27 03:55:24 localhost podman[100822]: 2025-11-27 08:55:24.034786079 +0000 UTC m=+0.124699790 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-type=git, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:55:24 localhost podman[100821]: 2025-11-27 08:55:24.038970931 +0000 UTC m=+0.132635113 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_id=tripleo_step3, distribution-scope=public, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, version=17.1.12, container_name=collectd, name=rhosp17/openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vcs-type=git, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:55:24 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:55:24 localhost podman[100822]: 2025-11-27 08:55:24.075506776 +0000 UTC m=+0.165420487 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, architecture=x86_64, vcs-type=git, distribution-scope=public, tcib_managed=true, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, managed_by=tripleo_ansible, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:55:24 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:55:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:55:36 localhost podman[100862]: 2025-11-27 08:55:36.02398916 +0000 UTC m=+0.088489895 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, version=17.1.12, vcs-type=git, distribution-scope=public, build-date=2025-11-18T22:49:46Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, container_name=metrics_qdr, io.buildah.version=1.41.4, managed_by=tripleo_ansible, release=1761123044, io.openshift.expose-services=, batch=17.1_20251118.1) Nov 27 03:55:36 localhost podman[100862]: 2025-11-27 08:55:36.218423206 +0000 UTC m=+0.282923911 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, name=rhosp17/openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, version=17.1.12, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, vendor=Red Hat, Inc., architecture=x86_64, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:55:36 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:55:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:55:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:55:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:55:41 localhost systemd[1]: tmp-crun.SFpliV.mount: Deactivated successfully. Nov 27 03:55:41 localhost podman[100892]: 2025-11-27 08:55:41.028151234 +0000 UTC m=+0.114696760 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.expose-services=, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, container_name=logrotate_crond, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, vcs-type=git, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:55:41 localhost podman[100893]: 2025-11-27 08:55:41.066866037 +0000 UTC m=+0.150877625 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, url=https://www.redhat.com, vcs-type=git, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, config_id=tripleo_step4, version=17.1.12, release=1761123044, build-date=2025-11-19T00:11:48Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:55:41 localhost podman[100891]: 2025-11-27 08:55:41.120696687 +0000 UTC m=+0.209424892 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, architecture=x86_64, io.buildah.version=1.41.4, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, com.redhat.component=openstack-ceilometer-ipmi-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, tcib_managed=true, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:55:41 localhost podman[100892]: 2025-11-27 08:55:41.142439682 +0000 UTC m=+0.228985208 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, architecture=x86_64, vcs-type=git, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, release=1761123044, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, tcib_managed=true, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:55:41 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:55:41 localhost podman[100893]: 2025-11-27 08:55:41.199791957 +0000 UTC m=+0.283803475 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, url=https://www.redhat.com, tcib_managed=true, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, architecture=x86_64, container_name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=) Nov 27 03:55:41 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:55:41 localhost podman[100891]: 2025-11-27 08:55:41.251792208 +0000 UTC m=+0.340520463 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, managed_by=tripleo_ansible, url=https://www.redhat.com, vendor=Red Hat, Inc., container_name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, distribution-scope=public, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.expose-services=, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:12:45Z, vcs-type=git) Nov 27 03:55:41 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:55:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:55:43 localhost systemd[1]: tmp-crun.up3j0V.mount: Deactivated successfully. Nov 27 03:55:43 localhost podman[100964]: 2025-11-27 08:55:43.9958281 +0000 UTC m=+0.091008581 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, distribution-scope=public, io.buildah.version=1.41.4, tcib_managed=true, architecture=x86_64, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:55:44 localhost podman[100964]: 2025-11-27 08:55:44.387793477 +0000 UTC m=+0.482973948 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, container_name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 03:55:44 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:55:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:55:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:55:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:55:47 localhost systemd[1]: tmp-crun.yT1p81.mount: Deactivated successfully. Nov 27 03:55:47 localhost podman[100989]: 2025-11-27 08:55:47.02411783 +0000 UTC m=+0.104475795 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:55:47 localhost podman[100987]: 2025-11-27 08:55:47.064621381 +0000 UTC m=+0.152384566 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, batch=17.1_20251118.1, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, maintainer=OpenStack TripleO Team, tcib_managed=true, io.openshift.expose-services=) Nov 27 03:55:47 localhost podman[100988]: 2025-11-27 08:55:47.113200969 +0000 UTC m=+0.198574539 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, distribution-scope=public, config_id=tripleo_step4, container_name=ovn_controller, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, release=1761123044, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vcs-type=git, architecture=x86_64) Nov 27 03:55:47 localhost podman[100988]: 2025-11-27 08:55:47.131819201 +0000 UTC m=+0.217192751 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.buildah.version=1.41.4, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, url=https://www.redhat.com, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc.) Nov 27 03:55:47 localhost podman[100988]: unhealthy Nov 27 03:55:47 localhost podman[100989]: 2025-11-27 08:55:47.139260272 +0000 UTC m=+0.219618237 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, config_id=tripleo_step5, distribution-scope=public, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044) Nov 27 03:55:47 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:55:47 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:55:47 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:55:47 localhost podman[100987]: 2025-11-27 08:55:47.183820051 +0000 UTC m=+0.271583236 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, release=1761123044, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, tcib_managed=true, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-type=git, managed_by=tripleo_ansible, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:55:47 localhost podman[100987]: unhealthy Nov 27 03:55:47 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:55:47 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:55:48 localhost systemd[1]: tmp-crun.t6Nnq0.mount: Deactivated successfully. Nov 27 03:55:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:55:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:55:54 localhost podman[101051]: 2025-11-27 08:55:54.973945369 +0000 UTC m=+0.069841422 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, version=17.1.12, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, config_id=tripleo_step3, io.buildah.version=1.41.4, batch=17.1_20251118.1, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, tcib_managed=true, com.redhat.component=openstack-collectd-container, release=1761123044, architecture=x86_64, build-date=2025-11-18T22:51:28Z) Nov 27 03:55:54 localhost podman[101051]: 2025-11-27 08:55:54.983724443 +0000 UTC m=+0.079620496 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.openshift.expose-services=, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, build-date=2025-11-18T22:51:28Z, io.buildah.version=1.41.4, vendor=Red Hat, Inc., container_name=collectd, vcs-type=git) Nov 27 03:55:54 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:55:55 localhost podman[101052]: 2025-11-27 08:55:55.037111651 +0000 UTC m=+0.129507269 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, io.openshift.expose-services=, vcs-type=git, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, tcib_managed=true, vendor=Red Hat, Inc., io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, distribution-scope=public) Nov 27 03:55:55 localhost podman[101052]: 2025-11-27 08:55:55.074872627 +0000 UTC m=+0.167268235 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, tcib_managed=true, distribution-scope=public, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step3, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, container_name=iscsid, url=https://www.redhat.com, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12) Nov 27 03:55:55 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:56:06 localhost systemd[1]: session-28.scope: Deactivated successfully. Nov 27 03:56:06 localhost systemd-logind[761]: Session 28 logged out. Waiting for processes to exit. Nov 27 03:56:06 localhost systemd[1]: session-28.scope: Consumed 7min 11.452s CPU time. Nov 27 03:56:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:56:06 localhost systemd-logind[761]: Removed session 28. Nov 27 03:56:06 localhost podman[101166]: 2025-11-27 08:56:06.43881817 +0000 UTC m=+0.091563368 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.openshift.expose-services=, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:56:06 localhost podman[101166]: 2025-11-27 08:56:06.647191361 +0000 UTC m=+0.299936589 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., architecture=x86_64, release=1761123044, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd, io.buildah.version=1.41.4, container_name=metrics_qdr, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=) Nov 27 03:56:06 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:56:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:56:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:56:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:56:12 localhost systemd[1]: tmp-crun.bx4ExJ.mount: Deactivated successfully. Nov 27 03:56:12 localhost podman[101196]: 2025-11-27 08:56:12.011342612 +0000 UTC m=+0.101881514 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., release=1761123044, name=rhosp17/openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, url=https://www.redhat.com, managed_by=tripleo_ansible, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, batch=17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 03:56:12 localhost podman[101197]: 2025-11-27 08:56:11.981024346 +0000 UTC m=+0.071847926 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, name=rhosp17/openstack-cron, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vcs-type=git, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., version=17.1.12, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, release=1761123044, io.openshift.expose-services=) Nov 27 03:56:12 localhost podman[101197]: 2025-11-27 08:56:12.063943379 +0000 UTC m=+0.154766989 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, config_id=tripleo_step4, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, maintainer=OpenStack TripleO Team, container_name=logrotate_crond) Nov 27 03:56:12 localhost podman[101198]: 2025-11-27 08:56:12.07769129 +0000 UTC m=+0.162924249 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, config_id=tripleo_step4, tcib_managed=true, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute) Nov 27 03:56:12 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:56:12 localhost podman[101196]: 2025-11-27 08:56:12.086415334 +0000 UTC m=+0.176954256 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=tripleo_step4, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, tcib_managed=true, release=1761123044, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 03:56:12 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:56:12 localhost podman[101198]: 2025-11-27 08:56:12.109108706 +0000 UTC m=+0.194341635 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, distribution-scope=public, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, container_name=ceilometer_agent_compute, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container) Nov 27 03:56:12 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:56:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:56:14 localhost systemd[1]: tmp-crun.RxDr0k.mount: Deactivated successfully. Nov 27 03:56:14 localhost podman[101269]: 2025-11-27 08:56:14.992701618 +0000 UTC m=+0.091522275 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, maintainer=OpenStack TripleO Team, distribution-scope=public, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, release=1761123044, tcib_managed=true, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, vcs-type=git) Nov 27 03:56:15 localhost podman[101269]: 2025-11-27 08:56:15.315996446 +0000 UTC m=+0.414817103 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, name=rhosp17/openstack-nova-compute, config_id=tripleo_step4, url=https://www.redhat.com, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, release=1761123044, vcs-type=git, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:56:15 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:56:16 localhost systemd[1]: Stopping User Manager for UID 1003... Nov 27 03:56:16 localhost systemd[36158]: Activating special unit Exit the Session... Nov 27 03:56:16 localhost systemd[36158]: Removed slice User Background Tasks Slice. Nov 27 03:56:16 localhost systemd[36158]: Stopped target Main User Target. Nov 27 03:56:16 localhost systemd[36158]: Stopped target Basic System. Nov 27 03:56:16 localhost systemd[36158]: Stopped target Paths. Nov 27 03:56:16 localhost systemd[36158]: Stopped target Sockets. Nov 27 03:56:16 localhost systemd[36158]: Stopped target Timers. Nov 27 03:56:16 localhost systemd[36158]: Stopped Mark boot as successful after the user session has run 2 minutes. Nov 27 03:56:16 localhost systemd[36158]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 03:56:16 localhost systemd[36158]: Closed D-Bus User Message Bus Socket. Nov 27 03:56:16 localhost systemd[36158]: Stopped Create User's Volatile Files and Directories. Nov 27 03:56:16 localhost systemd[36158]: Removed slice User Application Slice. Nov 27 03:56:16 localhost systemd[36158]: Reached target Shutdown. Nov 27 03:56:16 localhost systemd[36158]: Finished Exit the Session. Nov 27 03:56:16 localhost systemd[36158]: Reached target Exit the Session. Nov 27 03:56:16 localhost systemd[1]: user@1003.service: Deactivated successfully. Nov 27 03:56:16 localhost systemd[1]: Stopped User Manager for UID 1003. Nov 27 03:56:16 localhost systemd[1]: user@1003.service: Consumed 5.973s CPU time, read 0B from disk, written 7.0K to disk. Nov 27 03:56:16 localhost systemd[1]: Stopping User Runtime Directory /run/user/1003... Nov 27 03:56:16 localhost systemd[1]: run-user-1003.mount: Deactivated successfully. Nov 27 03:56:16 localhost systemd[1]: user-runtime-dir@1003.service: Deactivated successfully. Nov 27 03:56:16 localhost systemd[1]: Stopped User Runtime Directory /run/user/1003. Nov 27 03:56:16 localhost systemd[1]: Removed slice User Slice of UID 1003. Nov 27 03:56:16 localhost systemd[1]: user-1003.slice: Consumed 7min 17.456s CPU time. Nov 27 03:56:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:56:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:56:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:56:17 localhost podman[101294]: 2025-11-27 08:56:17.989960692 +0000 UTC m=+0.083552932 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, version=17.1.12, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, vendor=Red Hat, Inc., tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.buildah.version=1.41.4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, release=1761123044) Nov 27 03:56:18 localhost podman[101294]: 2025-11-27 08:56:18.028953732 +0000 UTC m=+0.122545972 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, distribution-scope=public, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, architecture=x86_64, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:56:18 localhost podman[101294]: unhealthy Nov 27 03:56:18 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:56:18 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:56:18 localhost podman[101296]: 2025-11-27 08:56:18.040980896 +0000 UTC m=+0.128847872 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, container_name=nova_compute, distribution-scope=public, config_id=tripleo_step5, architecture=x86_64, release=1761123044, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:56:18 localhost podman[101295]: 2025-11-27 08:56:18.097983832 +0000 UTC m=+0.187478041 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, managed_by=tripleo_ansible, build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, config_id=tripleo_step4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, io.openshift.expose-services=, version=17.1.12, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 03:56:18 localhost podman[101295]: 2025-11-27 08:56:18.115094932 +0000 UTC m=+0.204589141 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.openshift.expose-services=, url=https://www.redhat.com, vcs-type=git, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4) Nov 27 03:56:18 localhost podman[101295]: unhealthy Nov 27 03:56:18 localhost podman[101296]: 2025-11-27 08:56:18.122906423 +0000 UTC m=+0.210773389 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, release=1761123044, config_id=tripleo_step5, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, distribution-scope=public, vendor=Red Hat, Inc., batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4) Nov 27 03:56:18 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:56:18 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:56:18 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:56:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:56:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:56:25 localhost podman[101361]: 2025-11-27 08:56:25.988332608 +0000 UTC m=+0.080743045 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, url=https://www.redhat.com, vcs-type=git, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, config_id=tripleo_step3, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:56:26 localhost podman[101361]: 2025-11-27 08:56:26.000874456 +0000 UTC m=+0.093284853 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, release=1761123044, build-date=2025-11-18T23:44:13Z, vcs-type=git, tcib_managed=true, com.redhat.component=openstack-iscsid-container, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, container_name=iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_id=tripleo_step3, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4) Nov 27 03:56:26 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:56:26 localhost podman[101360]: 2025-11-27 08:56:26.095807393 +0000 UTC m=+0.190647716 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, managed_by=tripleo_ansible, io.buildah.version=1.41.4, tcib_managed=true, batch=17.1_20251118.1, container_name=collectd, name=rhosp17/openstack-collectd, config_id=tripleo_step3, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, version=17.1.12, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:56:26 localhost podman[101360]: 2025-11-27 08:56:26.10498266 +0000 UTC m=+0.199822963 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.expose-services=, tcib_managed=true, vcs-type=git, com.redhat.component=openstack-collectd-container, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, container_name=collectd, url=https://www.redhat.com, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-collectd) Nov 27 03:56:26 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:56:34 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:56:34 localhost recover_tripleo_nova_virtqemud[101401]: 63639 Nov 27 03:56:34 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:56:34 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:56:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:56:36 localhost systemd[1]: tmp-crun.5WL7Xp.mount: Deactivated successfully. Nov 27 03:56:36 localhost podman[101402]: 2025-11-27 08:56:36.992776036 +0000 UTC m=+0.092174093 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, build-date=2025-11-18T22:49:46Z, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, batch=17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., managed_by=tripleo_ansible, version=17.1.12, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:56:37 localhost podman[101402]: 2025-11-27 08:56:37.1789651 +0000 UTC m=+0.278363137 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.expose-services=, container_name=metrics_qdr, version=17.1.12, distribution-scope=public, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.buildah.version=1.41.4, tcib_managed=true, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd) Nov 27 03:56:37 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:56:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:56:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:56:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:56:42 localhost podman[101433]: 2025-11-27 08:56:42.997116638 +0000 UTC m=+0.086174533 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, version=17.1.12, distribution-scope=public, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, release=1761123044, build-date=2025-11-19T00:11:48Z, vendor=Red Hat, Inc., io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, vcs-type=git, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 03:56:43 localhost podman[101433]: 2025-11-27 08:56:43.029220082 +0000 UTC m=+0.118277947 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, vcs-type=git, io.openshift.expose-services=, release=1761123044, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, architecture=x86_64, build-date=2025-11-19T00:11:48Z) Nov 27 03:56:43 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:56:43 localhost podman[101431]: 2025-11-27 08:56:43.047453213 +0000 UTC m=+0.140218297 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, version=17.1.12, distribution-scope=public, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, config_id=tripleo_step4, architecture=x86_64, vcs-type=git) Nov 27 03:56:43 localhost podman[101431]: 2025-11-27 08:56:43.08075592 +0000 UTC m=+0.173520984 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, batch=17.1_20251118.1, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, architecture=x86_64, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, vcs-type=git, container_name=ceilometer_agent_ipmi, distribution-scope=public) Nov 27 03:56:43 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:56:43 localhost podman[101432]: 2025-11-27 08:56:43.093309998 +0000 UTC m=+0.184916551 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_id=tripleo_step4, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, tcib_managed=true, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, release=1761123044, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:32Z, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, architecture=x86_64, io.buildah.version=1.41.4) Nov 27 03:56:43 localhost podman[101432]: 2025-11-27 08:56:43.126002469 +0000 UTC m=+0.217609022 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, architecture=x86_64, name=rhosp17/openstack-cron) Nov 27 03:56:43 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:56:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:56:45 localhost systemd[1]: tmp-crun.xFXUpN.mount: Deactivated successfully. Nov 27 03:56:45 localhost podman[101502]: 2025-11-27 08:56:45.998636796 +0000 UTC m=+0.091683540 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, release=1761123044, managed_by=tripleo_ansible, url=https://www.redhat.com, container_name=nova_migration_target, vcs-type=git, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute) Nov 27 03:56:46 localhost podman[101502]: 2025-11-27 08:56:46.393357597 +0000 UTC m=+0.486404351 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_step4, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., vcs-type=git, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:56:46 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:56:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:56:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:56:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:56:49 localhost podman[101525]: 2025-11-27 08:56:49.031224611 +0000 UTC m=+0.121137413 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, distribution-scope=public, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, vendor=Red Hat, Inc., name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.buildah.version=1.41.4) Nov 27 03:56:49 localhost systemd[1]: tmp-crun.EqqKAm.mount: Deactivated successfully. Nov 27 03:56:49 localhost podman[101527]: 2025-11-27 08:56:49.085560604 +0000 UTC m=+0.170213464 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, name=rhosp17/openstack-nova-compute, version=17.1.12, container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.expose-services=, architecture=x86_64, config_id=tripleo_step5, batch=17.1_20251118.1) Nov 27 03:56:49 localhost podman[101527]: 2025-11-27 08:56:49.107329381 +0000 UTC m=+0.191982291 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, version=17.1.12, batch=17.1_20251118.1, distribution-scope=public, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 03:56:49 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:56:49 localhost podman[101526]: 2025-11-27 08:56:49.055772263 +0000 UTC m=+0.142717165 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, config_id=tripleo_step4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, container_name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, release=1761123044, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:56:49 localhost podman[101526]: 2025-11-27 08:56:49.19081643 +0000 UTC m=+0.277761342 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, managed_by=tripleo_ansible, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:56:49 localhost podman[101526]: unhealthy Nov 27 03:56:49 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:56:49 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:56:49 localhost podman[101525]: 2025-11-27 08:56:49.2097484 +0000 UTC m=+0.299661242 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, config_id=tripleo_step4, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible) Nov 27 03:56:49 localhost podman[101525]: unhealthy Nov 27 03:56:49 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:56:49 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:56:50 localhost systemd[1]: tmp-crun.tECzgE.mount: Deactivated successfully. Nov 27 03:56:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:56:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:56:56 localhost podman[101585]: 2025-11-27 08:56:56.992931071 +0000 UTC m=+0.082147813 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, com.redhat.component=openstack-iscsid-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, config_id=tripleo_step3, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, tcib_managed=true, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., version=17.1.12, architecture=x86_64) Nov 27 03:56:57 localhost podman[101585]: 2025-11-27 08:56:57.000939757 +0000 UTC m=+0.090156499 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, config_id=tripleo_step3, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:56:57 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:56:57 localhost systemd[1]: tmp-crun.22Sn8j.mount: Deactivated successfully. Nov 27 03:56:57 localhost podman[101584]: 2025-11-27 08:56:57.060696576 +0000 UTC m=+0.151408369 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, version=17.1.12, distribution-scope=public, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vcs-type=git, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., config_id=tripleo_step3) Nov 27 03:56:57 localhost podman[101584]: 2025-11-27 08:56:57.073016179 +0000 UTC m=+0.163728002 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, distribution-scope=public, io.buildah.version=1.41.4, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-collectd-container, release=1761123044, url=https://www.redhat.com, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:56:57 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:57:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:57:08 localhost systemd[1]: tmp-crun.ueeAkb.mount: Deactivated successfully. Nov 27 03:57:08 localhost podman[101699]: 2025-11-27 08:57:08.014716406 +0000 UTC m=+0.114381521 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, distribution-scope=public, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com) Nov 27 03:57:08 localhost podman[101699]: 2025-11-27 08:57:08.207049896 +0000 UTC m=+0.306714961 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, url=https://www.redhat.com, container_name=metrics_qdr, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, config_id=tripleo_step1, maintainer=OpenStack TripleO Team, tcib_managed=true) Nov 27 03:57:08 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:57:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:57:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:57:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:57:13 localhost systemd[1]: tmp-crun.0Jhpc5.mount: Deactivated successfully. Nov 27 03:57:14 localhost podman[101728]: 2025-11-27 08:57:14.005713548 +0000 UTC m=+0.100683202 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, build-date=2025-11-19T00:12:45Z, io.openshift.expose-services=, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, container_name=ceilometer_agent_ipmi, architecture=x86_64, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4) Nov 27 03:57:14 localhost podman[101728]: 2025-11-27 08:57:14.034993556 +0000 UTC m=+0.129963180 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, batch=17.1_20251118.1, config_id=tripleo_step4, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.openshift.expose-services=, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, io.buildah.version=1.41.4, vendor=Red Hat, Inc., url=https://www.redhat.com) Nov 27 03:57:14 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:57:14 localhost podman[101730]: 2025-11-27 08:57:14.055033537 +0000 UTC m=+0.144751990 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, config_id=tripleo_step4, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:11:48Z, name=rhosp17/openstack-ceilometer-compute, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, architecture=x86_64, io.buildah.version=1.41.4, distribution-scope=public) Nov 27 03:57:14 localhost podman[101730]: 2025-11-27 08:57:14.091787467 +0000 UTC m=+0.181505930 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, config_id=tripleo_step4, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, io.buildah.version=1.41.4, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, architecture=x86_64, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=) Nov 27 03:57:14 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:57:14 localhost podman[101729]: 2025-11-27 08:57:14.095747423 +0000 UTC m=+0.187873901 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, build-date=2025-11-18T22:49:32Z, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_id=tripleo_step4, architecture=x86_64, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, container_name=logrotate_crond, distribution-scope=public, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:57:14 localhost podman[101729]: 2025-11-27 08:57:14.204508912 +0000 UTC m=+0.296635350 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, version=17.1.12, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, config_id=tripleo_step4, distribution-scope=public, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, architecture=x86_64, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, tcib_managed=true, vcs-type=git, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:57:14 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:57:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:57:16 localhost podman[101802]: 2025-11-27 08:57:16.985994284 +0000 UTC m=+0.080013346 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, distribution-scope=public, vendor=Red Hat, Inc.) Nov 27 03:57:17 localhost podman[101802]: 2025-11-27 08:57:17.378106055 +0000 UTC m=+0.472125097 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, tcib_managed=true, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, container_name=nova_migration_target, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, url=https://www.redhat.com, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute) Nov 27 03:57:17 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:57:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:57:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:57:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:57:20 localhost systemd[1]: tmp-crun.JAH0i3.mount: Deactivated successfully. Nov 27 03:57:20 localhost podman[101827]: 2025-11-27 08:57:20.017698416 +0000 UTC m=+0.098636437 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, name=rhosp17/openstack-nova-compute, release=1761123044, vcs-type=git, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, container_name=nova_compute, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, tcib_managed=true, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:57:20 localhost podman[101827]: 2025-11-27 08:57:20.054918339 +0000 UTC m=+0.135856350 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step5, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_compute, version=17.1.12, release=1761123044, batch=17.1_20251118.1) Nov 27 03:57:20 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:57:20 localhost podman[101826]: 2025-11-27 08:57:20.069307306 +0000 UTC m=+0.152151359 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, config_id=tripleo_step4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 03:57:20 localhost podman[101826]: 2025-11-27 08:57:20.08686459 +0000 UTC m=+0.169708643 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, container_name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, vendor=Red Hat, Inc., build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_id=tripleo_step4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:57:20 localhost podman[101826]: unhealthy Nov 27 03:57:20 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:57:20 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:57:20 localhost podman[101825]: 2025-11-27 08:57:20.157916803 +0000 UTC m=+0.243877279 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, tcib_managed=true, container_name=ovn_metadata_agent, release=1761123044, build-date=2025-11-19T00:14:25Z, architecture=x86_64, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, config_id=tripleo_step4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 03:57:20 localhost podman[101825]: 2025-11-27 08:57:20.17786498 +0000 UTC m=+0.263825446 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, config_id=tripleo_step4, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 03:57:20 localhost podman[101825]: unhealthy Nov 27 03:57:20 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:57:20 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:57:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:57:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:57:27 localhost systemd[1]: tmp-crun.jNZvCT.mount: Deactivated successfully. Nov 27 03:57:28 localhost podman[101891]: 2025-11-27 08:57:28.004358518 +0000 UTC m=+0.098109264 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, config_id=tripleo_step3, release=1761123044, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, version=17.1.12, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, io.buildah.version=1.41.4) Nov 27 03:57:28 localhost podman[101891]: 2025-11-27 08:57:28.014009337 +0000 UTC m=+0.107760053 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., version=17.1.12, distribution-scope=public, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-type=git, io.openshift.expose-services=, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd) Nov 27 03:57:28 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:57:28 localhost podman[101892]: 2025-11-27 08:57:28.094486265 +0000 UTC m=+0.181041566 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, managed_by=tripleo_ansible, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, version=17.1.12, tcib_managed=true, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, container_name=iscsid, io.openshift.expose-services=) Nov 27 03:57:28 localhost podman[101892]: 2025-11-27 08:57:28.131010699 +0000 UTC m=+0.217566030 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, maintainer=OpenStack TripleO Team, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, io.buildah.version=1.41.4, distribution-scope=public, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, architecture=x86_64, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=) Nov 27 03:57:28 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:57:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:57:38 localhost podman[101928]: 2025-11-27 08:57:38.98297 +0000 UTC m=+0.082155333 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, config_id=tripleo_step1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:57:39 localhost podman[101928]: 2025-11-27 08:57:39.178911718 +0000 UTC m=+0.278097011 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, architecture=x86_64, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, batch=17.1_20251118.1, vendor=Red Hat, Inc., container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, name=rhosp17/openstack-qdrouterd, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.buildah.version=1.41.4, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:57:39 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:57:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:57:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:57:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:57:44 localhost systemd[1]: tmp-crun.OEWd2t.mount: Deactivated successfully. Nov 27 03:57:45 localhost podman[101958]: 2025-11-27 08:57:45.005984966 +0000 UTC m=+0.097164138 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.buildah.version=1.41.4, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, version=17.1.12, io.openshift.expose-services=, url=https://www.redhat.com, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, maintainer=OpenStack TripleO Team) Nov 27 03:57:45 localhost podman[101958]: 2025-11-27 08:57:45.014857425 +0000 UTC m=+0.106036627 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, tcib_managed=true, version=17.1.12, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, architecture=x86_64, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-type=git) Nov 27 03:57:45 localhost podman[101957]: 2025-11-27 08:57:44.971628171 +0000 UTC m=+0.067581911 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-ipmi-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, release=1761123044, container_name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, io.openshift.expose-services=, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, config_id=tripleo_step4, version=17.1.12, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 03:57:45 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:57:45 localhost podman[101957]: 2025-11-27 08:57:45.054532773 +0000 UTC m=+0.150486553 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, config_id=tripleo_step4, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 03:57:45 localhost podman[101959]: 2025-11-27 08:57:45.0670229 +0000 UTC m=+0.151404529 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, release=1761123044, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, tcib_managed=true, vendor=Red Hat, Inc.) Nov 27 03:57:45 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:57:45 localhost podman[101959]: 2025-11-27 08:57:45.095956689 +0000 UTC m=+0.180338318 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, batch=17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute, release=1761123044, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:57:45 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:57:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:57:47 localhost systemd[1]: tmp-crun.HoVSHO.mount: Deactivated successfully. Nov 27 03:57:47 localhost podman[102025]: 2025-11-27 08:57:47.995955244 +0000 UTC m=+0.095900314 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, architecture=x86_64, config_id=tripleo_step4, vcs-type=git, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, com.redhat.component=openstack-nova-compute-container, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:57:48 localhost podman[102025]: 2025-11-27 08:57:48.359145776 +0000 UTC m=+0.459090896 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, distribution-scope=public, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, config_id=tripleo_step4, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=) Nov 27 03:57:48 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:57:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:57:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:57:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:57:50 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:57:50 localhost recover_tripleo_nova_virtqemud[102062]: 63639 Nov 27 03:57:50 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:57:50 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:57:51 localhost podman[102048]: 2025-11-27 08:57:51.004288896 +0000 UTC m=+0.096824588 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, distribution-scope=public, release=1761123044, tcib_managed=true, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-type=git) Nov 27 03:57:51 localhost podman[102048]: 2025-11-27 08:57:51.045515226 +0000 UTC m=+0.138050948 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, vendor=Red Hat, Inc., config_id=tripleo_step4, managed_by=tripleo_ansible, container_name=ovn_metadata_agent, distribution-scope=public, tcib_managed=true, architecture=x86_64, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn) Nov 27 03:57:51 localhost podman[102048]: unhealthy Nov 27 03:57:51 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:57:51 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:57:51 localhost podman[102050]: 2025-11-27 08:57:51.061748144 +0000 UTC m=+0.146484397 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step5, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-type=git, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, version=17.1.12) Nov 27 03:57:51 localhost podman[102050]: 2025-11-27 08:57:51.102042249 +0000 UTC m=+0.186778552 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, container_name=nova_compute, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, architecture=x86_64, io.buildah.version=1.41.4, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, build-date=2025-11-19T00:36:58Z, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:57:51 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:57:51 localhost podman[102049]: 2025-11-27 08:57:51.117756082 +0000 UTC m=+0.207380466 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, io.openshift.expose-services=, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, release=1761123044, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4) Nov 27 03:57:51 localhost podman[102049]: 2025-11-27 08:57:51.157712068 +0000 UTC m=+0.247336482 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, tcib_managed=true, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12) Nov 27 03:57:51 localhost podman[102049]: unhealthy Nov 27 03:57:51 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:57:51 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 4200.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:57:51 localhost systemd[1]: tmp-crun.9vvI7d.mount: Deactivated successfully. Nov 27 03:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 03:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 4200.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 03:57:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:57:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:57:58 localhost podman[102115]: 2025-11-27 08:57:58.980641939 +0000 UTC m=+0.068150696 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, release=1761123044, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, distribution-scope=public, build-date=2025-11-18T23:44:13Z, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, vendor=Red Hat, Inc., architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.openshift.expose-services=, container_name=iscsid, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:57:58 localhost podman[102115]: 2025-11-27 08:57:58.990930746 +0000 UTC m=+0.078439473 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, architecture=x86_64, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, config_id=tripleo_step3, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, io.openshift.expose-services=, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, vendor=Red Hat, Inc., name=rhosp17/openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 03:57:59 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:57:59 localhost podman[102114]: 2025-11-27 08:57:59.044234502 +0000 UTC m=+0.134740280 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.buildah.version=1.41.4, config_id=tripleo_step3, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, version=17.1.12, tcib_managed=true, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, distribution-scope=public) Nov 27 03:57:59 localhost podman[102114]: 2025-11-27 08:57:59.0787218 +0000 UTC m=+0.169227588 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vcs-type=git, distribution-scope=public, release=1761123044, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:51:28Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:57:59 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:58:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:58:10 localhost podman[102229]: 2025-11-27 08:58:10.001908631 +0000 UTC m=+0.092670156 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, tcib_managed=true, version=17.1.12, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, distribution-scope=public, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:58:10 localhost podman[102229]: 2025-11-27 08:58:10.203871431 +0000 UTC m=+0.294632936 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.expose-services=, config_id=tripleo_step1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, container_name=metrics_qdr) Nov 27 03:58:10 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:58:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:58:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:58:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:58:15 localhost systemd[1]: tmp-crun.xqJXNP.mount: Deactivated successfully. Nov 27 03:58:15 localhost podman[102259]: 2025-11-27 08:58:15.988787894 +0000 UTC m=+0.082868714 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, tcib_managed=true, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-ipmi, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12) Nov 27 03:58:16 localhost podman[102260]: 2025-11-27 08:58:16.042696615 +0000 UTC m=+0.134449222 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.component=openstack-cron-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, config_id=tripleo_step4, container_name=logrotate_crond, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:32Z, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, vendor=Red Hat, Inc.) Nov 27 03:58:16 localhost podman[102259]: 2025-11-27 08:58:16.071731107 +0000 UTC m=+0.165811927 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vcs-type=git, release=1761123044, container_name=ceilometer_agent_ipmi, version=17.1.12, architecture=x86_64, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:58:16 localhost podman[102260]: 2025-11-27 08:58:16.079012264 +0000 UTC m=+0.170764821 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, build-date=2025-11-18T22:49:32Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, managed_by=tripleo_ansible, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:58:16 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:58:16 localhost podman[102261]: 2025-11-27 08:58:16.104943742 +0000 UTC m=+0.193680398 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, version=17.1.12, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, name=rhosp17/openstack-ceilometer-compute, tcib_managed=true, architecture=x86_64, config_id=tripleo_step4, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 03:58:16 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:58:16 localhost podman[102261]: 2025-11-27 08:58:16.137257612 +0000 UTC m=+0.225994268 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, build-date=2025-11-19T00:11:48Z, io.openshift.expose-services=, architecture=x86_64, distribution-scope=public, batch=17.1_20251118.1, config_id=tripleo_step4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com) Nov 27 03:58:16 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:58:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:58:18 localhost podman[102333]: 2025-11-27 08:58:18.981186147 +0000 UTC m=+0.075962387 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, name=rhosp17/openstack-nova-compute, vcs-type=git, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc.) Nov 27 03:58:19 localhost podman[102333]: 2025-11-27 08:58:19.353051322 +0000 UTC m=+0.447827592 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:58:19 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:58:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:58:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:58:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:58:22 localhost podman[102355]: 2025-11-27 08:58:22.017578425 +0000 UTC m=+0.105456372 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, container_name=ovn_controller, release=1761123044, architecture=x86_64, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, distribution-scope=public, version=17.1.12, tcib_managed=true, url=https://www.redhat.com, vcs-type=git, com.redhat.component=openstack-ovn-controller-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 03:58:22 localhost podman[102354]: 2025-11-27 08:58:22.03599648 +0000 UTC m=+0.131162053 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.expose-services=, url=https://www.redhat.com, vendor=Red Hat, Inc., config_id=tripleo_step4, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, tcib_managed=true, container_name=ovn_metadata_agent, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 03:58:22 localhost podman[102355]: 2025-11-27 08:58:22.040102171 +0000 UTC m=+0.127980098 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, release=1761123044) Nov 27 03:58:22 localhost podman[102355]: unhealthy Nov 27 03:58:22 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:58:22 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:58:22 localhost podman[102354]: 2025-11-27 08:58:22.059873324 +0000 UTC m=+0.155038927 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, architecture=x86_64) Nov 27 03:58:22 localhost podman[102354]: unhealthy Nov 27 03:58:22 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:58:22 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:58:22 localhost podman[102359]: 2025-11-27 08:58:22.114870065 +0000 UTC m=+0.199602067 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, container_name=nova_compute, batch=17.1_20251118.1, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, vendor=Red Hat, Inc., io.openshift.expose-services=, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044, config_id=tripleo_step5, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 03:58:22 localhost podman[102359]: 2025-11-27 08:58:22.172067125 +0000 UTC m=+0.256799067 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_id=tripleo_step5, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com, release=1761123044, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-type=git, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, version=17.1.12, name=rhosp17/openstack-nova-compute) Nov 27 03:58:22 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:58:23 localhost systemd[1]: tmp-crun.dfNyRf.mount: Deactivated successfully. Nov 27 03:58:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:58:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:58:29 localhost podman[102419]: 2025-11-27 08:58:29.997094413 +0000 UTC m=+0.088531395 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, container_name=collectd, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, vendor=Red Hat, Inc., url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, tcib_managed=true, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, maintainer=OpenStack TripleO Team, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 03:58:30 localhost podman[102419]: 2025-11-27 08:58:30.00622919 +0000 UTC m=+0.097666162 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-collectd, vcs-type=git, batch=17.1_20251118.1, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, io.buildah.version=1.41.4, config_id=tripleo_step3, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true) Nov 27 03:58:30 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:58:30 localhost podman[102420]: 2025-11-27 08:58:30.094352273 +0000 UTC m=+0.182861586 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, tcib_managed=true, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vendor=Red Hat, Inc., managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.buildah.version=1.41.4, distribution-scope=public, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:58:30 localhost podman[102420]: 2025-11-27 08:58:30.105844422 +0000 UTC m=+0.194353695 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vendor=Red Hat, Inc., batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, release=1761123044, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, distribution-scope=public, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=iscsid, config_id=tripleo_step3, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:58:30 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:58:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:58:40 localhost podman[102458]: 2025-11-27 08:58:40.981583683 +0000 UTC m=+0.081691291 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, batch=17.1_20251118.1, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, release=1761123044, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, container_name=metrics_qdr, config_id=tripleo_step1) Nov 27 03:58:41 localhost podman[102458]: 2025-11-27 08:58:41.174009556 +0000 UTC m=+0.274117124 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, distribution-scope=public, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, build-date=2025-11-18T22:49:46Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, vcs-type=git, architecture=x86_64, managed_by=tripleo_ansible, io.openshift.expose-services=, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1) Nov 27 03:58:41 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:58:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:58:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:58:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:58:46 localhost systemd[1]: tmp-crun.2xJpkw.mount: Deactivated successfully. Nov 27 03:58:47 localhost podman[102488]: 2025-11-27 08:58:47.002023519 +0000 UTC m=+0.091158836 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, url=https://www.redhat.com, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 03:58:47 localhost podman[102488]: 2025-11-27 08:58:47.035026128 +0000 UTC m=+0.124161435 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, name=rhosp17/openstack-cron, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, url=https://www.redhat.com, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, distribution-scope=public, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron) Nov 27 03:58:47 localhost podman[102487]: 2025-11-27 08:58:47.049559829 +0000 UTC m=+0.138922122 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.buildah.version=1.41.4, batch=17.1_20251118.1, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, container_name=ceilometer_agent_ipmi, distribution-scope=public, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:58:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:58:47 localhost podman[102487]: 2025-11-27 08:58:47.104934541 +0000 UTC m=+0.194296804 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, build-date=2025-11-19T00:12:45Z, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., distribution-scope=public, batch=17.1_20251118.1, architecture=x86_64, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 03:58:47 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:58:47 localhost podman[102489]: 2025-11-27 08:58:47.106671888 +0000 UTC m=+0.190689707 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ceilometer-compute-container, io.buildah.version=1.41.4, container_name=ceilometer_agent_compute, tcib_managed=true, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, version=17.1.12) Nov 27 03:58:47 localhost podman[102489]: 2025-11-27 08:58:47.189931641 +0000 UTC m=+0.273949470 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, config_id=tripleo_step4, io.openshift.expose-services=, container_name=ceilometer_agent_compute, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, io.buildah.version=1.41.4, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 03:58:47 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:58:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:58:49 localhost systemd[1]: tmp-crun.7kuutA.mount: Deactivated successfully. Nov 27 03:58:49 localhost podman[102558]: 2025-11-27 08:58:49.993084606 +0000 UTC m=+0.092937904 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, release=1761123044, version=17.1.12, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, container_name=nova_migration_target, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com) Nov 27 03:58:50 localhost podman[102558]: 2025-11-27 08:58:50.373889612 +0000 UTC m=+0.473742830 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, container_name=nova_migration_target, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, config_id=tripleo_step4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 03:58:50 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:58:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:58:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:58:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:58:52 localhost podman[102584]: 2025-11-27 08:58:52.988202393 +0000 UTC m=+0.082806961 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible) Nov 27 03:58:53 localhost podman[102584]: 2025-11-27 08:58:53.001663646 +0000 UTC m=+0.096268194 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, version=17.1.12, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.openshift.expose-services=, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, batch=17.1_20251118.1, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-type=git, architecture=x86_64, distribution-scope=public, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 03:58:53 localhost podman[102584]: unhealthy Nov 27 03:58:53 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:58:53 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:58:53 localhost podman[102586]: 2025-11-27 08:58:53.042995989 +0000 UTC m=+0.131887823 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, config_id=tripleo_step5, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, release=1761123044, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:58:53 localhost systemd[1]: tmp-crun.Fy3Jc9.mount: Deactivated successfully. Nov 27 03:58:53 localhost podman[102585]: 2025-11-27 08:58:53.096473599 +0000 UTC m=+0.187930792 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, vendor=Red Hat, Inc., config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, distribution-scope=public, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, release=1761123044, url=https://www.redhat.com, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:58:53 localhost podman[102585]: 2025-11-27 08:58:53.110501057 +0000 UTC m=+0.201958230 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_id=tripleo_step4, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, release=1761123044, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, container_name=ovn_controller, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, distribution-scope=public, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git) Nov 27 03:58:53 localhost podman[102585]: unhealthy Nov 27 03:58:53 localhost podman[102586]: 2025-11-27 08:58:53.120468266 +0000 UTC m=+0.209360130 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, vendor=Red Hat, Inc., config_id=tripleo_step5, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-type=git, release=1761123044, name=rhosp17/openstack-nova-compute) Nov 27 03:58:53 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:58:53 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:58:53 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:59:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:59:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:59:00 localhost systemd[1]: tmp-crun.GtdDsK.mount: Deactivated successfully. Nov 27 03:59:01 localhost podman[102651]: 2025-11-27 08:59:01.005612252 +0000 UTC m=+0.100638691 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, maintainer=OpenStack TripleO Team, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.expose-services=, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, distribution-scope=public, com.redhat.component=openstack-collectd-container, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, architecture=x86_64, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, vcs-type=git) Nov 27 03:59:01 localhost podman[102651]: 2025-11-27 08:59:01.015253163 +0000 UTC m=+0.110279572 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, container_name=collectd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, managed_by=tripleo_ansible) Nov 27 03:59:01 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:59:01 localhost podman[102652]: 2025-11-27 08:59:01.099645885 +0000 UTC m=+0.189927956 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, vcs-type=git, batch=17.1_20251118.1, container_name=iscsid, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., release=1761123044, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 03:59:01 localhost podman[102652]: 2025-11-27 08:59:01.133411305 +0000 UTC m=+0.223693386 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, architecture=x86_64, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vendor=Red Hat, Inc., io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, tcib_managed=true, name=rhosp17/openstack-iscsid, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, url=https://www.redhat.com, io.buildah.version=1.41.4, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:59:01 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:59:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:59:11 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 03:59:11 localhost recover_tripleo_nova_virtqemud[102772]: 63639 Nov 27 03:59:11 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 03:59:11 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 03:59:11 localhost podman[102770]: 2025-11-27 08:59:11.997586665 +0000 UTC m=+0.092217465 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, container_name=metrics_qdr, io.buildah.version=1.41.4, batch=17.1_20251118.1) Nov 27 03:59:12 localhost podman[102770]: 2025-11-27 08:59:12.191321063 +0000 UTC m=+0.285951813 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, version=17.1.12, tcib_managed=true, vcs-type=git, container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.buildah.version=1.41.4, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-qdrouterd, release=1761123044, batch=17.1_20251118.1, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, distribution-scope=public, config_id=tripleo_step1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 03:59:12 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:59:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:59:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:59:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:59:18 localhost podman[102801]: 2025-11-27 08:59:18.002464503 +0000 UTC m=+0.093727056 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., build-date=2025-11-19T00:12:45Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, distribution-scope=public, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 03:59:18 localhost systemd[1]: tmp-crun.bjVaog.mount: Deactivated successfully. Nov 27 03:59:18 localhost podman[102801]: 2025-11-27 08:59:18.064271508 +0000 UTC m=+0.155534041 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-type=git, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, build-date=2025-11-19T00:12:45Z, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, config_id=tripleo_step4, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:59:18 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:59:18 localhost podman[102803]: 2025-11-27 08:59:18.065223173 +0000 UTC m=+0.151470641 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, vcs-type=git, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, release=1761123044, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:59:18 localhost podman[102802]: 2025-11-27 08:59:18.156187462 +0000 UTC m=+0.245358008 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-type=git, container_name=logrotate_crond, name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:59:18 localhost podman[102802]: 2025-11-27 08:59:18.193983381 +0000 UTC m=+0.283153887 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, architecture=x86_64, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, distribution-scope=public, io.openshift.expose-services=, io.buildah.version=1.41.4, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, container_name=logrotate_crond, build-date=2025-11-18T22:49:32Z, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc.) Nov 27 03:59:18 localhost podman[102803]: 2025-11-27 08:59:18.200937798 +0000 UTC m=+0.287185286 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, container_name=ceilometer_agent_compute, tcib_managed=true, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, version=17.1.12, distribution-scope=public, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team) Nov 27 03:59:18 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:59:18 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:59:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:59:20 localhost systemd[1]: tmp-crun.a81Ngo.mount: Deactivated successfully. Nov 27 03:59:20 localhost podman[102872]: 2025-11-27 08:59:20.991257498 +0000 UTC m=+0.083546921 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_id=tripleo_step4, version=17.1.12, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, url=https://www.redhat.com) Nov 27 03:59:21 localhost podman[102872]: 2025-11-27 08:59:21.368684193 +0000 UTC m=+0.460973616 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, version=17.1.12) Nov 27 03:59:21 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:59:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:59:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:59:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:59:23 localhost systemd[1]: tmp-crun.fdqa1L.mount: Deactivated successfully. Nov 27 03:59:24 localhost podman[102896]: 2025-11-27 08:59:24.003750762 +0000 UTC m=+0.093163920 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, tcib_managed=true, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, vcs-type=git, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 03:59:24 localhost podman[102896]: 2025-11-27 08:59:24.045941888 +0000 UTC m=+0.135355056 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, architecture=x86_64, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, vcs-type=git, batch=17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-ovn-controller-container) Nov 27 03:59:24 localhost podman[102896]: unhealthy Nov 27 03:59:24 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:59:24 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:59:24 localhost podman[102897]: 2025-11-27 08:59:24.050129151 +0000 UTC m=+0.134349770 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, tcib_managed=true, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_id=tripleo_step5, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, vendor=Red Hat, Inc.) Nov 27 03:59:24 localhost podman[102895]: 2025-11-27 08:59:24.108430581 +0000 UTC m=+0.199480293 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, vcs-type=git, version=17.1.12, batch=17.1_20251118.1, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, container_name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:59:24 localhost podman[102897]: 2025-11-27 08:59:24.130817004 +0000 UTC m=+0.215037553 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_id=tripleo_step5, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, distribution-scope=public, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true) Nov 27 03:59:24 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:59:24 localhost podman[102895]: 2025-11-27 08:59:24.145995613 +0000 UTC m=+0.237045325 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.buildah.version=1.41.4, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, name=rhosp17/openstack-neutron-metadata-agent-ovn, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, distribution-scope=public, container_name=ovn_metadata_agent, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 03:59:24 localhost podman[102895]: unhealthy Nov 27 03:59:24 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:59:24 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:59:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 03:59:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 03:59:31 localhost systemd[1]: tmp-crun.wmNHR7.mount: Deactivated successfully. Nov 27 03:59:31 localhost podman[102963]: 2025-11-27 08:59:31.99213595 +0000 UTC m=+0.083491189 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, version=17.1.12, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, config_id=tripleo_step3, distribution-scope=public, maintainer=OpenStack TripleO Team, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, io.openshift.expose-services=, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}) Nov 27 03:59:32 localhost podman[102963]: 2025-11-27 08:59:32.001173254 +0000 UTC m=+0.092528493 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, tcib_managed=true, io.buildah.version=1.41.4, version=17.1.12, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, config_id=tripleo_step3) Nov 27 03:59:32 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 03:59:32 localhost podman[102962]: 2025-11-27 08:59:32.099183034 +0000 UTC m=+0.194372487 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, url=https://www.redhat.com, version=17.1.12, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, batch=17.1_20251118.1) Nov 27 03:59:32 localhost podman[102962]: 2025-11-27 08:59:32.112329807 +0000 UTC m=+0.207519250 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=openstack-collectd-container, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, config_id=tripleo_step3, batch=17.1_20251118.1, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, release=1761123044, tcib_managed=true, version=17.1.12, container_name=collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z) Nov 27 03:59:32 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 03:59:32 localhost systemd[1]: tmp-crun.lCJLlu.mount: Deactivated successfully. Nov 27 03:59:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 03:59:42 localhost systemd[1]: tmp-crun.8z4mBV.mount: Deactivated successfully. Nov 27 03:59:43 localhost podman[102998]: 2025-11-27 08:59:42.999729785 +0000 UTC m=+0.091516896 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.buildah.version=1.41.4, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-qdrouterd-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, tcib_managed=true, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd, vcs-type=git, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64) Nov 27 03:59:43 localhost podman[102998]: 2025-11-27 08:59:43.267103595 +0000 UTC m=+0.358890686 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, architecture=x86_64) Nov 27 03:59:43 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 03:59:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 03:59:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 03:59:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 03:59:49 localhost systemd[1]: tmp-crun.2eSmxb.mount: Deactivated successfully. Nov 27 03:59:49 localhost podman[103028]: 2025-11-27 08:59:49.039379328 +0000 UTC m=+0.093705225 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.openshift.expose-services=, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, release=1761123044, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, container_name=ceilometer_agent_ipmi, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, batch=17.1_20251118.1) Nov 27 03:59:49 localhost systemd[1]: tmp-crun.x9PGBY.mount: Deactivated successfully. Nov 27 03:59:49 localhost podman[103029]: 2025-11-27 08:59:49.084900664 +0000 UTC m=+0.139307723 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, vcs-type=git, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, release=1761123044, summary=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 03:59:49 localhost podman[103029]: 2025-11-27 08:59:49.094553713 +0000 UTC m=+0.148960782 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, container_name=logrotate_crond, name=rhosp17/openstack-cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, tcib_managed=true, release=1761123044, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:32Z, url=https://www.redhat.com, distribution-scope=public, vcs-type=git) Nov 27 03:59:49 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 03:59:49 localhost podman[103030]: 2025-11-27 08:59:49.142024522 +0000 UTC m=+0.193446211 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack TripleO Team, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, architecture=x86_64, com.redhat.component=openstack-ceilometer-compute-container, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com) Nov 27 03:59:49 localhost podman[103028]: 2025-11-27 08:59:49.195000969 +0000 UTC m=+0.249326876 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z) Nov 27 03:59:49 localhost podman[103030]: 2025-11-27 08:59:49.201993108 +0000 UTC m=+0.253414797 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, tcib_managed=true, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4) Nov 27 03:59:49 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 03:59:49 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 03:59:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 03:59:51 localhost podman[103100]: 2025-11-27 08:59:51.982708719 +0000 UTC m=+0.082080832 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, name=rhosp17/openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, release=1761123044, version=17.1.12, vendor=Red Hat, Inc., tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, batch=17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, distribution-scope=public) Nov 27 03:59:52 localhost podman[103100]: 2025-11-27 08:59:52.382021924 +0000 UTC m=+0.481394217 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, release=1761123044, version=17.1.12, build-date=2025-11-19T00:36:58Z, architecture=x86_64, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:59:52 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 03:59:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 03:59:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 03:59:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 03:59:54 localhost podman[103123]: 2025-11-27 08:59:54.990207529 +0000 UTC m=+0.084301632 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, architecture=x86_64, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vcs-type=git, build-date=2025-11-19T00:14:25Z) Nov 27 03:59:55 localhost podman[103123]: 2025-11-27 08:59:55.007938326 +0000 UTC m=+0.102032419 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., architecture=x86_64, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, release=1761123044, config_id=tripleo_step4, container_name=ovn_metadata_agent, version=17.1.12) Nov 27 03:59:55 localhost podman[103123]: unhealthy Nov 27 03:59:55 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:59:55 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 03:59:55 localhost systemd[1]: tmp-crun.s8iD61.mount: Deactivated successfully. Nov 27 03:59:55 localhost podman[103124]: 2025-11-27 08:59:55.109284165 +0000 UTC m=+0.197065268 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, vcs-type=git, url=https://www.redhat.com, vendor=Red Hat, Inc., managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, architecture=x86_64, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, tcib_managed=true, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, version=17.1.12, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4) Nov 27 03:59:55 localhost podman[103125]: 2025-11-27 08:59:55.159698823 +0000 UTC m=+0.246133449 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, version=17.1.12, architecture=x86_64, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 03:59:55 localhost podman[103124]: 2025-11-27 08:59:55.17178996 +0000 UTC m=+0.259571043 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, build-date=2025-11-18T23:34:05Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, release=1761123044, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, vendor=Red Hat, Inc., io.openshift.expose-services=, batch=17.1_20251118.1, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vcs-type=git) Nov 27 03:59:55 localhost podman[103124]: unhealthy Nov 27 03:59:55 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 03:59:55 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 03:59:55 localhost podman[103125]: 2025-11-27 08:59:55.22306837 +0000 UTC m=+0.309502996 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, managed_by=tripleo_ansible, container_name=nova_compute, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, batch=17.1_20251118.1, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 03:59:55 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 03:59:55 localhost systemd[1]: tmp-crun.GtATch.mount: Deactivated successfully. Nov 27 04:00:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:00:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:00:02 localhost podman[103190]: 2025-11-27 09:00:02.988079833 +0000 UTC m=+0.076992875 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, description=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, container_name=iscsid, architecture=x86_64) Nov 27 04:00:03 localhost podman[103190]: 2025-11-27 09:00:03.024916335 +0000 UTC m=+0.113829347 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.buildah.version=1.41.4, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, release=1761123044, managed_by=tripleo_ansible, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 04:00:03 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:00:03 localhost podman[103189]: 2025-11-27 09:00:03.041692286 +0000 UTC m=+0.132456018 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, distribution-scope=public, maintainer=OpenStack TripleO Team, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, config_id=tripleo_step3, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=collectd, io.openshift.expose-services=, url=https://www.redhat.com, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 04:00:03 localhost podman[103189]: 2025-11-27 09:00:03.052150638 +0000 UTC m=+0.142914370 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=collectd, vendor=Red Hat, Inc., url=https://www.redhat.com, version=17.1.12, architecture=x86_64, tcib_managed=true, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, config_id=tripleo_step3) Nov 27 04:00:03 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:00:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:00:13 localhost podman[103354]: 2025-11-27 09:00:13.990005052 +0000 UTC m=+0.084644171 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, config_id=tripleo_step1, url=https://www.redhat.com, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git, build-date=2025-11-18T22:49:46Z, version=17.1.12, release=1761123044, batch=17.1_20251118.1) Nov 27 04:00:14 localhost podman[103354]: 2025-11-27 09:00:14.20824092 +0000 UTC m=+0.302880079 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, distribution-scope=public, name=rhosp17/openstack-qdrouterd, architecture=x86_64, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, version=17.1.12, release=1761123044, com.redhat.component=openstack-qdrouterd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, managed_by=tripleo_ansible, io.openshift.expose-services=) Nov 27 04:00:14 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:00:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:00:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:00:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:00:20 localhost systemd[1]: tmp-crun.QGjmED.mount: Deactivated successfully. Nov 27 04:00:20 localhost podman[103383]: 2025-11-27 09:00:20.011656371 +0000 UTC m=+0.101810523 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-11-19T00:12:45Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, vcs-type=git, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, managed_by=tripleo_ansible) Nov 27 04:00:20 localhost podman[103384]: 2025-11-27 09:00:20.054733481 +0000 UTC m=+0.141379608 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, release=1761123044, url=https://www.redhat.com, architecture=x86_64, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=logrotate_crond, maintainer=OpenStack TripleO Team) Nov 27 04:00:20 localhost podman[103383]: 2025-11-27 09:00:20.074116293 +0000 UTC m=+0.164270445 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, tcib_managed=true, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, config_id=tripleo_step4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 04:00:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:00:20 localhost podman[103384]: 2025-11-27 09:00:20.089905919 +0000 UTC m=+0.176552026 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-18T22:49:32Z, com.redhat.component=openstack-cron-container, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, release=1761123044, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git) Nov 27 04:00:20 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:00:20 localhost podman[103385]: 2025-11-27 09:00:20.166470801 +0000 UTC m=+0.249014608 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, managed_by=tripleo_ansible, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, version=17.1.12, architecture=x86_64, container_name=ceilometer_agent_compute, vcs-type=git, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4) Nov 27 04:00:20 localhost podman[103385]: 2025-11-27 09:00:20.219969721 +0000 UTC m=+0.302513508 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, build-date=2025-11-19T00:11:48Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, io.openshift.expose-services=, io.buildah.version=1.41.4, config_id=tripleo_step4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12) Nov 27 04:00:20 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:00:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:00:23 localhost podman[103456]: 2025-11-27 09:00:23.01797715 +0000 UTC m=+0.120500377 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, container_name=nova_migration_target, vendor=Red Hat, Inc., release=1761123044) Nov 27 04:00:23 localhost podman[103456]: 2025-11-27 09:00:23.40008473 +0000 UTC m=+0.502607967 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, release=1761123044, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com) Nov 27 04:00:23 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:00:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:00:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:00:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:00:25 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:00:25 localhost recover_tripleo_nova_virtqemud[103498]: 63639 Nov 27 04:00:25 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:00:25 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:00:26 localhost podman[103479]: 2025-11-27 09:00:26.004064723 +0000 UTC m=+0.097019615 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:14:25Z, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.buildah.version=1.41.4) Nov 27 04:00:26 localhost podman[103481]: 2025-11-27 09:00:26.044725427 +0000 UTC m=+0.129573940 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, version=17.1.12, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, batch=17.1_20251118.1, managed_by=tripleo_ansible, config_id=tripleo_step5, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, container_name=nova_compute) Nov 27 04:00:26 localhost podman[103479]: 2025-11-27 09:00:26.04780878 +0000 UTC m=+0.140763662 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, release=1761123044, batch=17.1_20251118.1, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, managed_by=tripleo_ansible, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:00:26 localhost podman[103479]: unhealthy Nov 27 04:00:26 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:00:26 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:00:26 localhost podman[103480]: 2025-11-27 09:00:26.117964021 +0000 UTC m=+0.201637232 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, architecture=x86_64, vcs-type=git, version=17.1.12, tcib_managed=true, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1) Nov 27 04:00:26 localhost podman[103481]: 2025-11-27 09:00:26.125324928 +0000 UTC m=+0.210173441 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, version=17.1.12, batch=17.1_20251118.1, config_id=tripleo_step5, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 04:00:26 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:00:26 localhost podman[103480]: 2025-11-27 09:00:26.162940852 +0000 UTC m=+0.246614063 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, build-date=2025-11-18T23:34:05Z, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller) Nov 27 04:00:26 localhost podman[103480]: unhealthy Nov 27 04:00:26 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:00:26 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:00:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:00:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:00:34 localhost podman[103551]: 2025-11-27 09:00:34.003634001 +0000 UTC m=+0.100567439 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, version=17.1.12, distribution-scope=public, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, io.openshift.expose-services=) Nov 27 04:00:34 localhost systemd[1]: tmp-crun.Dfu2X5.mount: Deactivated successfully. Nov 27 04:00:34 localhost podman[103552]: 2025-11-27 09:00:34.049652491 +0000 UTC m=+0.143355742 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, config_id=tripleo_step3, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 04:00:34 localhost podman[103551]: 2025-11-27 09:00:34.067595604 +0000 UTC m=+0.164528982 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., io.openshift.expose-services=, architecture=x86_64, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, vcs-type=git, com.redhat.component=openstack-collectd-container, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, container_name=collectd) Nov 27 04:00:34 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:00:34 localhost podman[103552]: 2025-11-27 09:00:34.086987796 +0000 UTC m=+0.180691067 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, vendor=Red Hat, Inc., architecture=x86_64, managed_by=tripleo_ansible, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.expose-services=, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, build-date=2025-11-18T23:44:13Z, distribution-scope=public, com.redhat.component=openstack-iscsid-container, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, maintainer=OpenStack TripleO Team) Nov 27 04:00:34 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:00:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:00:44 localhost podman[103590]: 2025-11-27 09:00:44.982348287 +0000 UTC m=+0.082233906 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, name=rhosp17/openstack-qdrouterd, config_id=tripleo_step1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z) Nov 27 04:00:45 localhost podman[103590]: 2025-11-27 09:00:45.198203661 +0000 UTC m=+0.298089230 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step1, vendor=Red Hat, Inc., batch=17.1_20251118.1, version=17.1.12, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, container_name=metrics_qdr, release=1761123044) Nov 27 04:00:45 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:00:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:00:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:00:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:00:50 localhost systemd[1]: tmp-crun.HDhtwS.mount: Deactivated successfully. Nov 27 04:00:51 localhost podman[103621]: 2025-11-27 09:00:51.00196046 +0000 UTC m=+0.092290267 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, maintainer=OpenStack TripleO Team, vcs-type=git, io.openshift.expose-services=, config_id=tripleo_step4, version=17.1.12, url=https://www.redhat.com, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, distribution-scope=public, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, release=1761123044, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 04:00:51 localhost podman[103621]: 2025-11-27 09:00:51.029759089 +0000 UTC m=+0.120088906 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, io.buildah.version=1.41.4, build-date=2025-11-19T00:11:48Z, architecture=x86_64, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, batch=17.1_20251118.1, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, io.openshift.expose-services=, managed_by=tripleo_ansible, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 04:00:51 localhost systemd[1]: tmp-crun.NfcWxo.mount: Deactivated successfully. Nov 27 04:00:51 localhost podman[103619]: 2025-11-27 09:00:51.040999372 +0000 UTC m=+0.136623731 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, version=17.1.12, config_id=tripleo_step4, managed_by=tripleo_ansible, vcs-type=git, container_name=ceilometer_agent_ipmi, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 04:00:51 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:00:51 localhost podman[103620]: 2025-11-27 09:00:51.09219427 +0000 UTC m=+0.184498609 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, name=rhosp17/openstack-cron, release=1761123044, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_id=tripleo_step4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, container_name=logrotate_crond, build-date=2025-11-18T22:49:32Z, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, distribution-scope=public) Nov 27 04:00:51 localhost podman[103620]: 2025-11-27 09:00:51.104867632 +0000 UTC m=+0.197172011 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, container_name=logrotate_crond, vendor=Red Hat, Inc., url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, name=rhosp17/openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, tcib_managed=true, build-date=2025-11-18T22:49:32Z, release=1761123044, architecture=x86_64, com.redhat.component=openstack-cron-container) Nov 27 04:00:51 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:00:51 localhost podman[103619]: 2025-11-27 09:00:51.148739024 +0000 UTC m=+0.244363423 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, vcs-type=git, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:12:45Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, tcib_managed=true) Nov 27 04:00:51 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:00:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:00:53 localhost podman[103694]: 2025-11-27 09:00:53.984760735 +0000 UTC m=+0.080910220 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, managed_by=tripleo_ansible, config_id=tripleo_step4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc.) Nov 27 04:00:54 localhost podman[103694]: 2025-11-27 09:00:54.358049679 +0000 UTC m=+0.454199194 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public) Nov 27 04:00:54 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:00:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:00:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:00:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:00:56 localhost systemd[1]: tmp-crun.qRVgeS.mount: Deactivated successfully. Nov 27 04:00:57 localhost podman[103720]: 2025-11-27 09:00:57.001929616 +0000 UTC m=+0.089413780 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step5, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, release=1761123044, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, distribution-scope=public, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true) Nov 27 04:00:57 localhost systemd[1]: tmp-crun.Jc6Yia.mount: Deactivated successfully. Nov 27 04:00:57 localhost podman[103718]: 2025-11-27 09:00:57.051628194 +0000 UTC m=+0.147227216 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, managed_by=tripleo_ansible, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 04:00:57 localhost podman[103720]: 2025-11-27 09:00:57.054027089 +0000 UTC m=+0.141511293 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, version=17.1.12, container_name=nova_compute, url=https://www.redhat.com, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, tcib_managed=true, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, release=1761123044) Nov 27 04:00:57 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:00:57 localhost podman[103719]: 2025-11-27 09:00:57.100950312 +0000 UTC m=+0.192407012 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, release=1761123044, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, url=https://www.redhat.com, distribution-scope=public, version=17.1.12, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:00:57 localhost podman[103718]: 2025-11-27 09:00:57.119860012 +0000 UTC m=+0.215459044 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, vcs-type=git, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, batch=17.1_20251118.1, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, version=17.1.12) Nov 27 04:00:57 localhost podman[103718]: unhealthy Nov 27 04:00:57 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:00:57 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:00:57 localhost podman[103719]: 2025-11-27 09:00:57.170864715 +0000 UTC m=+0.262321425 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, managed_by=tripleo_ansible, config_id=tripleo_step4, vcs-type=git, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, tcib_managed=true, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public) Nov 27 04:00:57 localhost podman[103719]: unhealthy Nov 27 04:00:57 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:00:57 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:01:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:01:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:01:04 localhost podman[103811]: 2025-11-27 09:01:04.989044268 +0000 UTC m=+0.084382633 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, com.redhat.component=openstack-iscsid-container, tcib_managed=true, io.buildah.version=1.41.4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, build-date=2025-11-18T23:44:13Z, container_name=iscsid, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, architecture=x86_64, config_id=tripleo_step3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, maintainer=OpenStack TripleO Team) Nov 27 04:01:05 localhost podman[103811]: 2025-11-27 09:01:05.027960556 +0000 UTC m=+0.123298931 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.buildah.version=1.41.4, container_name=iscsid, config_id=tripleo_step3, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, build-date=2025-11-18T23:44:13Z, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, distribution-scope=public, managed_by=tripleo_ansible, release=1761123044, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 04:01:05 localhost podman[103810]: 2025-11-27 09:01:05.058097048 +0000 UTC m=+0.156170517 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, container_name=collectd, managed_by=tripleo_ansible, tcib_managed=true, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, distribution-scope=public, com.redhat.component=openstack-collectd-container, batch=17.1_20251118.1, io.openshift.expose-services=, name=rhosp17/openstack-collectd, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd) Nov 27 04:01:05 localhost podman[103810]: 2025-11-27 09:01:05.071104418 +0000 UTC m=+0.169177887 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vcs-type=git, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, build-date=2025-11-18T22:51:28Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, name=rhosp17/openstack-collectd, version=17.1.12, release=1761123044) Nov 27 04:01:05 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:01:05 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:01:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:01:15 localhost podman[103927]: 2025-11-27 09:01:15.987512946 +0000 UTC m=+0.080781527 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step1, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, release=1761123044, com.redhat.component=openstack-qdrouterd-container, tcib_managed=true, build-date=2025-11-18T22:49:46Z, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, url=https://www.redhat.com, io.buildah.version=1.41.4, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team) Nov 27 04:01:16 localhost podman[103927]: 2025-11-27 09:01:16.214016777 +0000 UTC m=+0.307285408 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, config_id=tripleo_step1, batch=17.1_20251118.1) Nov 27 04:01:16 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:01:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:01:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:01:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:01:21 localhost systemd[1]: tmp-crun.CMWFPW.mount: Deactivated successfully. Nov 27 04:01:21 localhost podman[103955]: 2025-11-27 09:01:21.991144499 +0000 UTC m=+0.089098450 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, batch=17.1_20251118.1, container_name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-ipmi, managed_by=tripleo_ansible, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi) Nov 27 04:01:21 localhost podman[103956]: 2025-11-27 09:01:21.996664608 +0000 UTC m=+0.087450966 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, vcs-type=git, com.redhat.component=openstack-cron-container, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, vendor=Red Hat, Inc., distribution-scope=public, version=17.1.12, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 cron, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, container_name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:01:22 localhost podman[103956]: 2025-11-27 09:01:22.031911597 +0000 UTC m=+0.122697975 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, container_name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, io.buildah.version=1.41.4, release=1761123044, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:01:22 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:01:22 localhost podman[103959]: 2025-11-27 09:01:22.049051579 +0000 UTC m=+0.137036102 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ceilometer_agent_compute, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, architecture=x86_64, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, tcib_managed=true, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, io.buildah.version=1.41.4) Nov 27 04:01:22 localhost podman[103955]: 2025-11-27 09:01:22.065211414 +0000 UTC m=+0.163165365 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.buildah.version=1.41.4, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, config_id=tripleo_step4, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 04:01:22 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:01:22 localhost podman[103959]: 2025-11-27 09:01:22.103779973 +0000 UTC m=+0.191764516 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, io.buildah.version=1.41.4, vcs-type=git, managed_by=tripleo_ansible, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, io.openshift.expose-services=, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, tcib_managed=true, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, distribution-scope=public, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:01:22 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:01:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:01:24 localhost podman[104026]: 2025-11-27 09:01:24.986653837 +0000 UTC m=+0.082366710 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, container_name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, architecture=x86_64, batch=17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team) Nov 27 04:01:25 localhost podman[104026]: 2025-11-27 09:01:25.388307834 +0000 UTC m=+0.484020657 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, io.buildah.version=1.41.4, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, url=https://www.redhat.com, io.openshift.expose-services=, release=1761123044, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, version=17.1.12, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team) Nov 27 04:01:25 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:01:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:01:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:01:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:01:27 localhost podman[104053]: 2025-11-27 09:01:27.980658612 +0000 UTC m=+0.067779586 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, tcib_managed=true, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, version=17.1.12, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, container_name=nova_compute, architecture=x86_64, distribution-scope=public) Nov 27 04:01:28 localhost podman[104051]: 2025-11-27 09:01:28.042621992 +0000 UTC m=+0.138267766 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., config_id=tripleo_step4, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, version=17.1.12, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, architecture=x86_64) Nov 27 04:01:28 localhost podman[104053]: 2025-11-27 09:01:28.05702911 +0000 UTC m=+0.144150054 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, batch=17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, architecture=x86_64, release=1761123044, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, version=17.1.12, container_name=nova_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:01:28 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:01:28 localhost podman[104051]: 2025-11-27 09:01:28.08714457 +0000 UTC m=+0.182790344 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, batch=17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, architecture=x86_64, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, release=1761123044, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z) Nov 27 04:01:28 localhost podman[104051]: unhealthy Nov 27 04:01:28 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:01:28 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:01:28 localhost podman[104052]: 2025-11-27 09:01:28.143509139 +0000 UTC m=+0.235020191 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, managed_by=tripleo_ansible, vcs-type=git, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, tcib_managed=true, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vendor=Red Hat, Inc., config_id=tripleo_step4) Nov 27 04:01:28 localhost podman[104052]: 2025-11-27 09:01:28.156087577 +0000 UTC m=+0.247598619 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, architecture=x86_64, managed_by=tripleo_ansible, version=17.1.12, url=https://www.redhat.com, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, distribution-scope=public, vendor=Red Hat, Inc., release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git) Nov 27 04:01:28 localhost podman[104052]: unhealthy Nov 27 04:01:28 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:01:28 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:01:34 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:01:34 localhost recover_tripleo_nova_virtqemud[104115]: 63639 Nov 27 04:01:34 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:01:34 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:01:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:01:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:01:35 localhost systemd[1]: tmp-crun.Byt4pk.mount: Deactivated successfully. Nov 27 04:01:35 localhost podman[104116]: 2025-11-27 09:01:35.997836756 +0000 UTC m=+0.090119568 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, name=rhosp17/openstack-collectd, distribution-scope=public, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=) Nov 27 04:01:36 localhost podman[104116]: 2025-11-27 09:01:36.034698909 +0000 UTC m=+0.126981741 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, maintainer=OpenStack TripleO Team, release=1761123044, managed_by=tripleo_ansible, architecture=x86_64, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vendor=Red Hat, Inc., version=17.1.12, io.buildah.version=1.41.4, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, config_id=tripleo_step3, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd) Nov 27 04:01:36 localhost podman[104117]: 2025-11-27 09:01:36.04699066 +0000 UTC m=+0.137139324 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, release=1761123044, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, url=https://www.redhat.com, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:44:13Z, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, distribution-scope=public, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., version=17.1.12, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, tcib_managed=true, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid) Nov 27 04:01:36 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:01:36 localhost podman[104117]: 2025-11-27 09:01:36.061873131 +0000 UTC m=+0.152021785 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, architecture=x86_64, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=iscsid, io.buildah.version=1.41.4, managed_by=tripleo_ansible, build-date=2025-11-18T23:44:13Z, distribution-scope=public, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid) Nov 27 04:01:36 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:01:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:01:46 localhost podman[104157]: 2025-11-27 09:01:46.986737286 +0000 UTC m=+0.086231193 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, architecture=x86_64, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, tcib_managed=true, vcs-type=git, name=rhosp17/openstack-qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, url=https://www.redhat.com, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:01:47 localhost podman[104157]: 2025-11-27 09:01:47.174338989 +0000 UTC m=+0.273832886 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.buildah.version=1.41.4, config_id=tripleo_step1, io.openshift.expose-services=, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, tcib_managed=true, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, container_name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044) Nov 27 04:01:47 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:01:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:01:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:01:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:01:52 localhost podman[104187]: 2025-11-27 09:01:52.990821631 +0000 UTC m=+0.084254210 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, url=https://www.redhat.com, build-date=2025-11-19T00:12:45Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, batch=17.1_20251118.1, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 04:01:53 localhost systemd[1]: tmp-crun.BmPaQg.mount: Deactivated successfully. Nov 27 04:01:53 localhost podman[104187]: 2025-11-27 09:01:53.047970581 +0000 UTC m=+0.141403170 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, release=1761123044, container_name=ceilometer_agent_ipmi, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:01:53 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:01:53 localhost podman[104189]: 2025-11-27 09:01:53.097641618 +0000 UTC m=+0.183636546 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, batch=17.1_20251118.1, url=https://www.redhat.com, vcs-type=git, build-date=2025-11-19T00:11:48Z, version=17.1.12, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, distribution-scope=public, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 04:01:53 localhost podman[104189]: 2025-11-27 09:01:53.146288909 +0000 UTC m=+0.232283807 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, release=1761123044, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, url=https://www.redhat.com, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, architecture=x86_64, build-date=2025-11-19T00:11:48Z, batch=17.1_20251118.1, io.openshift.expose-services=, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 04:01:53 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:01:53 localhost podman[104188]: 2025-11-27 09:01:53.049933233 +0000 UTC m=+0.141584103 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.expose-services=, url=https://www.redhat.com, name=rhosp17/openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, distribution-scope=public, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, io.buildah.version=1.41.4, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, release=1761123044) Nov 27 04:01:53 localhost podman[104188]: 2025-11-27 09:01:53.230350333 +0000 UTC m=+0.322001203 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, batch=17.1_20251118.1, architecture=x86_64, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, vcs-type=git, io.buildah.version=1.41.4, com.redhat.component=openstack-cron-container, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 cron, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, container_name=logrotate_crond, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true) Nov 27 04:01:53 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:01:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:01:55 localhost systemd[1]: tmp-crun.ICGrzp.mount: Deactivated successfully. Nov 27 04:01:55 localhost podman[104258]: 2025-11-27 09:01:55.991912978 +0000 UTC m=+0.091229787 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, release=1761123044, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, container_name=nova_migration_target, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, distribution-scope=public, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 04:01:56 localhost podman[104258]: 2025-11-27 09:01:56.358940054 +0000 UTC m=+0.458256813 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-type=git, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, architecture=x86_64, tcib_managed=true, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, build-date=2025-11-19T00:36:58Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:01:56 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:01:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:01:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:01:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:01:58 localhost systemd[1]: tmp-crun.xhuU67.mount: Deactivated successfully. Nov 27 04:01:59 localhost podman[104281]: 2025-11-27 09:01:59.003772436 +0000 UTC m=+0.095914895 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, distribution-scope=public, tcib_managed=true, batch=17.1_20251118.1, release=1761123044, config_id=tripleo_step4, url=https://www.redhat.com, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, build-date=2025-11-19T00:14:25Z) Nov 27 04:01:59 localhost podman[104281]: 2025-11-27 09:01:59.02287515 +0000 UTC m=+0.115017569 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vcs-type=git, tcib_managed=true, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, architecture=x86_64, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 04:01:59 localhost podman[104283]: 2025-11-27 09:01:59.056437704 +0000 UTC m=+0.142230701 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, container_name=nova_compute, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, architecture=x86_64, url=https://www.redhat.com, vendor=Red Hat, Inc., version=17.1.12, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, build-date=2025-11-19T00:36:58Z, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, io.buildah.version=1.41.4) Nov 27 04:01:59 localhost podman[104281]: unhealthy Nov 27 04:01:59 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:01:59 localhost podman[104283]: 2025-11-27 09:01:59.087915052 +0000 UTC m=+0.173708029 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, tcib_managed=true, url=https://www.redhat.com, version=17.1.12, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:01:59 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:01:59 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:01:59 localhost podman[104282]: 2025-11-27 09:01:59.146184361 +0000 UTC m=+0.236771298 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.openshift.expose-services=, com.redhat.component=openstack-ovn-controller-container, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, tcib_managed=true, managed_by=tripleo_ansible, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, architecture=x86_64, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller) Nov 27 04:01:59 localhost podman[104282]: 2025-11-27 09:01:59.191165003 +0000 UTC m=+0.281751920 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, vcs-type=git, build-date=2025-11-18T23:34:05Z, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, maintainer=OpenStack TripleO Team, version=17.1.12, url=https://www.redhat.com, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:01:59 localhost podman[104282]: unhealthy Nov 27 04:01:59 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:01:59 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:02:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:02:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:02:06 localhost podman[104343]: 2025-11-27 09:02:06.995711528 +0000 UTC m=+0.088206036 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.buildah.version=1.41.4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., managed_by=tripleo_ansible, container_name=collectd, name=rhosp17/openstack-collectd, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:51:28Z, distribution-scope=public, tcib_managed=true, vcs-type=git, batch=17.1_20251118.1) Nov 27 04:02:07 localhost podman[104343]: 2025-11-27 09:02:07.004510336 +0000 UTC m=+0.097004844 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, com.redhat.component=openstack-collectd-container, io.openshift.expose-services=, config_id=tripleo_step3, version=17.1.12, container_name=collectd, name=rhosp17/openstack-collectd, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, vendor=Red Hat, Inc., managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team) Nov 27 04:02:07 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:02:07 localhost podman[104344]: 2025-11-27 09:02:07.094871629 +0000 UTC m=+0.183498033 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, release=1761123044, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, container_name=iscsid, build-date=2025-11-18T23:44:13Z, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, distribution-scope=public, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, url=https://www.redhat.com, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, summary=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, batch=17.1_20251118.1, vendor=Red Hat, Inc.) Nov 27 04:02:07 localhost podman[104344]: 2025-11-27 09:02:07.108941838 +0000 UTC m=+0.197568242 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12, managed_by=tripleo_ansible, name=rhosp17/openstack-iscsid, container_name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:44:13Z) Nov 27 04:02:07 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:02:09 localhost systemd[1]: tmp-crun.HKgArd.mount: Deactivated successfully. Nov 27 04:02:09 localhost podman[104485]: 2025-11-27 09:02:09.802111231 +0000 UTC m=+0.094069674 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.openshift.expose-services=, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, name=rhceph, version=7, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, architecture=x86_64, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:02:09 localhost podman[104485]: 2025-11-27 09:02:09.926291117 +0000 UTC m=+0.218249520 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_BRANCH=main, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, distribution-scope=public) Nov 27 04:02:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:02:18 localhost podman[104630]: 2025-11-27 09:02:18.006219159 +0000 UTC m=+0.087784074 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, container_name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, release=1761123044, vendor=Red Hat, Inc.) Nov 27 04:02:18 localhost podman[104630]: 2025-11-27 09:02:18.19967475 +0000 UTC m=+0.281239695 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd, config_id=tripleo_step1, io.openshift.expose-services=, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, release=1761123044, build-date=2025-11-18T22:49:46Z, name=rhosp17/openstack-qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 04:02:18 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:02:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:02:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:02:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:02:24 localhost systemd[1]: tmp-crun.mYPdZc.mount: Deactivated successfully. Nov 27 04:02:24 localhost podman[104660]: 2025-11-27 09:02:24.073541158 +0000 UTC m=+0.156557247 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, tcib_managed=true, version=17.1.12, io.buildah.version=1.41.4, vendor=Red Hat, Inc., release=1761123044, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:02:24 localhost podman[104661]: 2025-11-27 09:02:24.117835272 +0000 UTC m=+0.197035819 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, config_id=tripleo_step4, io.buildah.version=1.41.4, version=17.1.12, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., managed_by=tripleo_ansible, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, vcs-type=git, tcib_managed=true, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 04:02:24 localhost podman[104659]: 2025-11-27 09:02:24.028983898 +0000 UTC m=+0.116785776 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, architecture=x86_64, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.expose-services=, version=17.1.12, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible) Nov 27 04:02:24 localhost podman[104661]: 2025-11-27 09:02:24.154098037 +0000 UTC m=+0.233298574 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-compute-container, vcs-type=git, release=1761123044, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, container_name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, managed_by=tripleo_ansible, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team) Nov 27 04:02:24 localhost podman[104659]: 2025-11-27 09:02:24.163976634 +0000 UTC m=+0.251778542 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, version=17.1.12, vcs-type=git, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, architecture=x86_64, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vendor=Red Hat, Inc., build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, io.buildah.version=1.41.4, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4) Nov 27 04:02:24 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:02:24 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:02:24 localhost podman[104660]: 2025-11-27 09:02:24.204573187 +0000 UTC m=+0.287589276 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, architecture=x86_64, com.redhat.component=openstack-cron-container, container_name=logrotate_crond, version=17.1.12, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, release=1761123044, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, batch=17.1_20251118.1, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:02:24 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:02:25 localhost systemd[1]: tmp-crun.4Ep9rn.mount: Deactivated successfully. Nov 27 04:02:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:02:26 localhost systemd[1]: tmp-crun.KNA5iE.mount: Deactivated successfully. Nov 27 04:02:27 localhost podman[104733]: 2025-11-27 09:02:26.999323678 +0000 UTC m=+0.091313680 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., io.openshift.expose-services=, architecture=x86_64, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, distribution-scope=public, build-date=2025-11-19T00:36:58Z) Nov 27 04:02:27 localhost podman[104733]: 2025-11-27 09:02:27.376065245 +0000 UTC m=+0.468055277 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_migration_target, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, distribution-scope=public, vcs-type=git, com.redhat.component=openstack-nova-compute-container, release=1761123044) Nov 27 04:02:27 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:02:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:02:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:02:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:02:30 localhost systemd[1]: tmp-crun.t5FwtI.mount: Deactivated successfully. Nov 27 04:02:30 localhost podman[104755]: 2025-11-27 09:02:30.031794831 +0000 UTC m=+0.126440437 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-type=git, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, distribution-scope=public, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, managed_by=tripleo_ansible, tcib_managed=true, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:02:30 localhost podman[104755]: 2025-11-27 09:02:30.071735516 +0000 UTC m=+0.166381072 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, build-date=2025-11-19T00:14:25Z, tcib_managed=true, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, architecture=x86_64, managed_by=tripleo_ansible, io.openshift.expose-services=, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, config_id=tripleo_step4, version=17.1.12) Nov 27 04:02:30 localhost podman[104755]: unhealthy Nov 27 04:02:30 localhost podman[104757]: 2025-11-27 09:02:30.082441045 +0000 UTC m=+0.171587783 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, batch=17.1_20251118.1, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, config_id=tripleo_step5, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, version=17.1.12, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc.) Nov 27 04:02:30 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:02:30 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:02:30 localhost podman[104757]: 2025-11-27 09:02:30.115611558 +0000 UTC m=+0.204758286 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, release=1761123044, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, distribution-scope=public, tcib_managed=true, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, version=17.1.12, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:02:30 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:02:30 localhost podman[104756]: 2025-11-27 09:02:30.13463669 +0000 UTC m=+0.226230053 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, io.openshift.expose-services=, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, distribution-scope=public, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, config_id=tripleo_step4, vcs-type=git, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:02:30 localhost podman[104756]: 2025-11-27 09:02:30.151946846 +0000 UTC m=+0.243540209 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, url=https://www.redhat.com, com.redhat.component=openstack-ovn-controller-container, tcib_managed=true, distribution-scope=public, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:02:30 localhost podman[104756]: unhealthy Nov 27 04:02:30 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:02:30 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:02:31 localhost systemd[1]: tmp-crun.sGgGBS.mount: Deactivated successfully. Nov 27 04:02:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:02:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:02:38 localhost systemd[1]: tmp-crun.Xp8DEZ.mount: Deactivated successfully. Nov 27 04:02:38 localhost podman[104823]: 2025-11-27 09:02:38.010714733 +0000 UTC m=+0.098779101 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, vcs-type=git, architecture=x86_64, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, name=rhosp17/openstack-iscsid, version=17.1.12, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4) Nov 27 04:02:38 localhost podman[104822]: 2025-11-27 09:02:38.052683084 +0000 UTC m=+0.143525267 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, config_id=tripleo_step3, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, managed_by=tripleo_ansible, distribution-scope=public, io.openshift.expose-services=, version=17.1.12, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, tcib_managed=true, build-date=2025-11-18T22:51:28Z) Nov 27 04:02:38 localhost podman[104822]: 2025-11-27 09:02:38.068950242 +0000 UTC m=+0.159792465 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, io.buildah.version=1.41.4, architecture=x86_64, io.openshift.expose-services=, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, com.redhat.component=openstack-collectd-container, vendor=Red Hat, Inc., build-date=2025-11-18T22:51:28Z, version=17.1.12, maintainer=OpenStack TripleO Team, container_name=collectd, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:02:38 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:02:38 localhost podman[104823]: 2025-11-27 09:02:38.121699852 +0000 UTC m=+0.209764220 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, tcib_managed=true, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, release=1761123044, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, version=17.1.12, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public) Nov 27 04:02:38 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:02:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:02:49 localhost systemd[1]: tmp-crun.YdqkbW.mount: Deactivated successfully. Nov 27 04:02:49 localhost podman[104861]: 2025-11-27 09:02:49.022377746 +0000 UTC m=+0.122106770 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, name=rhosp17/openstack-qdrouterd, tcib_managed=true, release=1761123044, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., io.buildah.version=1.41.4, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team) Nov 27 04:02:49 localhost podman[104861]: 2025-11-27 09:02:49.209796503 +0000 UTC m=+0.309525487 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, com.redhat.component=openstack-qdrouterd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, version=17.1.12, url=https://www.redhat.com, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public) Nov 27 04:02:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:02:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:02:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:02:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:02:55 localhost systemd[1]: tmp-crun.u2uwtk.mount: Deactivated successfully. Nov 27 04:02:55 localhost podman[104893]: 2025-11-27 09:02:55.032216426 +0000 UTC m=+0.115072920 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, release=1761123044, batch=17.1_20251118.1, vendor=Red Hat, Inc., distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-compute, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, com.redhat.component=openstack-ceilometer-compute-container, config_id=tripleo_step4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, version=17.1.12, architecture=x86_64, vcs-type=git, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 04:02:55 localhost podman[104891]: 2025-11-27 09:02:55.047190789 +0000 UTC m=+0.137111364 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, io.openshift.expose-services=, managed_by=tripleo_ansible, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, architecture=x86_64, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, batch=17.1_20251118.1, config_id=tripleo_step4, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:02:55 localhost podman[104893]: 2025-11-27 09:02:55.101892693 +0000 UTC m=+0.184749147 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, vendor=Red Hat, Inc., distribution-scope=public) Nov 27 04:02:55 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:02:55 localhost podman[104891]: 2025-11-27 09:02:55.126268569 +0000 UTC m=+0.216189074 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step4, release=1761123044, vcs-type=git, io.openshift.expose-services=, version=17.1.12, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, io.buildah.version=1.41.4) Nov 27 04:02:55 localhost podman[104892]: 2025-11-27 09:02:54.984365588 +0000 UTC m=+0.073944443 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, name=rhosp17/openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=logrotate_crond, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-cron-container, batch=17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, vcs-type=git, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4) Nov 27 04:02:55 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:02:55 localhost podman[104892]: 2025-11-27 09:02:55.172764591 +0000 UTC m=+0.262343426 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, url=https://www.redhat.com, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, distribution-scope=public, batch=17.1_20251118.1, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, version=17.1.12, container_name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, architecture=x86_64, build-date=2025-11-18T22:49:32Z, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:02:55 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:02:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:02:57 localhost systemd[1]: tmp-crun.MhdARB.mount: Deactivated successfully. Nov 27 04:02:57 localhost podman[104964]: 2025-11-27 09:02:57.992380082 +0000 UTC m=+0.087796616 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, batch=17.1_20251118.1, url=https://www.redhat.com, distribution-scope=public, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vcs-type=git, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, container_name=nova_migration_target, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vendor=Red Hat, Inc.) Nov 27 04:02:58 localhost podman[104964]: 2025-11-27 09:02:58.364101833 +0000 UTC m=+0.459518327 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, distribution-scope=public, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, config_id=tripleo_step4, maintainer=OpenStack TripleO Team) Nov 27 04:02:58 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:03:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:03:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:03:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:03:00 localhost podman[104987]: 2025-11-27 09:03:00.992706589 +0000 UTC m=+0.086608714 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vendor=Red Hat, Inc., architecture=x86_64, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, container_name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:03:01 localhost podman[104988]: 2025-11-27 09:03:00.972127824 +0000 UTC m=+0.065167906 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, architecture=x86_64, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:34:05Z, version=17.1.12, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, container_name=ovn_controller, release=1761123044) Nov 27 04:03:01 localhost podman[104989]: 2025-11-27 09:03:01.041822051 +0000 UTC m=+0.131647946 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, version=17.1.12, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, release=1761123044, container_name=nova_compute, architecture=x86_64, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc.) Nov 27 04:03:01 localhost podman[104988]: 2025-11-27 09:03:01.060930906 +0000 UTC m=+0.153970958 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, vcs-type=git, config_id=tripleo_step4, url=https://www.redhat.com, tcib_managed=true, architecture=x86_64, container_name=ovn_controller, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:03:01 localhost podman[104988]: unhealthy Nov 27 04:03:01 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:03:01 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:03:01 localhost podman[104989]: 2025-11-27 09:03:01.100971184 +0000 UTC m=+0.190797109 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, build-date=2025-11-19T00:36:58Z, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vendor=Red Hat, Inc., org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, config_id=tripleo_step5, url=https://www.redhat.com, container_name=nova_compute) Nov 27 04:03:01 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:03:01 localhost podman[104987]: 2025-11-27 09:03:01.127219232 +0000 UTC m=+0.221121357 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, container_name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, version=17.1.12, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=) Nov 27 04:03:01 localhost podman[104987]: unhealthy Nov 27 04:03:01 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:03:01 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:03:01 localhost systemd[1]: tmp-crun.yQfR2A.mount: Deactivated successfully. Nov 27 04:03:04 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:03:04 localhost recover_tripleo_nova_virtqemud[105051]: 63639 Nov 27 04:03:04 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:03:04 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:03:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:03:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:03:08 localhost podman[105052]: 2025-11-27 09:03:08.992480323 +0000 UTC m=+0.089241064 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vendor=Red Hat, Inc., container_name=collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, distribution-scope=public, url=https://www.redhat.com, architecture=x86_64, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, io.openshift.expose-services=, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, com.redhat.component=openstack-collectd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, summary=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-collectd, batch=17.1_20251118.1) Nov 27 04:03:09 localhost podman[105052]: 2025-11-27 09:03:09.002159515 +0000 UTC m=+0.098920176 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, vendor=Red Hat, Inc., config_id=tripleo_step3, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, vcs-type=git, managed_by=tripleo_ansible, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4) Nov 27 04:03:09 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:03:09 localhost podman[105053]: 2025-11-27 09:03:09.087446711 +0000 UTC m=+0.183186234 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, io.buildah.version=1.41.4, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_id=tripleo_step3, architecture=x86_64, distribution-scope=public, version=17.1.12, managed_by=tripleo_ansible, container_name=iscsid, com.redhat.component=openstack-iscsid-container, vcs-type=git, release=1761123044, io.openshift.expose-services=) Nov 27 04:03:09 localhost podman[105053]: 2025-11-27 09:03:09.125023724 +0000 UTC m=+0.220763237 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, architecture=x86_64, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, io.buildah.version=1.41.4, config_id=tripleo_step3, distribution-scope=public, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, tcib_managed=true, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.component=openstack-iscsid-container, name=rhosp17/openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:03:09 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:03:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:03:19 localhost podman[105167]: 2025-11-27 09:03:19.992434741 +0000 UTC m=+0.080124739 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, vcs-type=git, tcib_managed=true, version=17.1.12, config_id=tripleo_step1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, distribution-scope=public, release=1761123044, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.openshift.expose-services=, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, container_name=metrics_qdr, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team) Nov 27 04:03:20 localhost podman[105167]: 2025-11-27 09:03:20.192392687 +0000 UTC m=+0.280082595 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, container_name=metrics_qdr, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, com.redhat.component=openstack-qdrouterd-container, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:03:20 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:03:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:03:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:03:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:03:26 localhost systemd[1]: tmp-crun.iQJXQ3.mount: Deactivated successfully. Nov 27 04:03:26 localhost podman[105197]: 2025-11-27 09:03:26.015252102 +0000 UTC m=+0.106902630 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, version=17.1.12, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, architecture=x86_64, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_ipmi, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, com.redhat.component=openstack-ceilometer-ipmi-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, batch=17.1_20251118.1) Nov 27 04:03:26 localhost podman[105198]: 2025-11-27 09:03:26.051095127 +0000 UTC m=+0.140775743 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, config_id=tripleo_step4, container_name=logrotate_crond, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vendor=Red Hat, Inc., managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, release=1761123044, build-date=2025-11-18T22:49:32Z) Nov 27 04:03:26 localhost podman[105198]: 2025-11-27 09:03:26.063863111 +0000 UTC m=+0.153543747 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, release=1761123044, tcib_managed=true, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, batch=17.1_20251118.1, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.buildah.version=1.41.4) Nov 27 04:03:26 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:03:26 localhost podman[105197]: 2025-11-27 09:03:26.150172035 +0000 UTC m=+0.241822563 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, io.buildah.version=1.41.4, distribution-scope=public, version=17.1.12, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, architecture=x86_64, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, container_name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 04:03:26 localhost podman[105199]: 2025-11-27 09:03:26.166733702 +0000 UTC m=+0.252035729 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, build-date=2025-11-19T00:11:48Z, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., url=https://www.redhat.com, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, managed_by=tripleo_ansible, io.buildah.version=1.41.4, distribution-scope=public, architecture=x86_64, batch=17.1_20251118.1, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, release=1761123044, maintainer=OpenStack TripleO Team) Nov 27 04:03:26 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:03:26 localhost podman[105199]: 2025-11-27 09:03:26.204959282 +0000 UTC m=+0.290261299 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, name=rhosp17/openstack-ceilometer-compute, vendor=Red Hat, Inc., container_name=ceilometer_agent_compute, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ceilometer-compute-container, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, build-date=2025-11-19T00:11:48Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, url=https://www.redhat.com, architecture=x86_64, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute) Nov 27 04:03:26 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:03:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:03:28 localhost systemd[1]: tmp-crun.p40X35.mount: Deactivated successfully. Nov 27 04:03:28 localhost podman[105270]: 2025-11-27 09:03:28.991966622 +0000 UTC m=+0.081954548 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vcs-type=git, config_id=tripleo_step4, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, tcib_managed=true, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, version=17.1.12, managed_by=tripleo_ansible) Nov 27 04:03:29 localhost podman[105270]: 2025-11-27 09:03:29.400158126 +0000 UTC m=+0.490146042 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vcs-type=git, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute) Nov 27 04:03:29 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:03:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:03:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:03:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:03:31 localhost podman[105291]: 2025-11-27 09:03:31.986730878 +0000 UTC m=+0.081447854 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, vcs-type=git, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, container_name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, tcib_managed=true, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, io.openshift.expose-services=, vendor=Red Hat, Inc., architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 04:03:32 localhost podman[105291]: 2025-11-27 09:03:32.024703851 +0000 UTC m=+0.119420797 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, release=1761123044, architecture=x86_64, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, container_name=ovn_metadata_agent, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:03:32 localhost podman[105291]: unhealthy Nov 27 04:03:32 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:03:32 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:03:32 localhost systemd[1]: tmp-crun.o0qNKW.mount: Deactivated successfully. Nov 27 04:03:32 localhost podman[105293]: 2025-11-27 09:03:32.1059741 +0000 UTC m=+0.194844079 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, url=https://www.redhat.com, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, batch=17.1_20251118.1, io.buildah.version=1.41.4, vendor=Red Hat, Inc., release=1761123044, config_id=tripleo_step5, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, com.redhat.component=openstack-nova-compute-container) Nov 27 04:03:32 localhost podman[105292]: 2025-11-27 09:03:32.149658367 +0000 UTC m=+0.240483418 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, managed_by=tripleo_ansible, tcib_managed=true, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, vcs-type=git, version=17.1.12, release=1761123044) Nov 27 04:03:32 localhost podman[105293]: 2025-11-27 09:03:32.161957508 +0000 UTC m=+0.250827457 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, managed_by=tripleo_ansible, batch=17.1_20251118.1, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, container_name=nova_compute, release=1761123044, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, vcs-type=git, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:03:32 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:03:32 localhost podman[105292]: 2025-11-27 09:03:32.189653034 +0000 UTC m=+0.280478115 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, build-date=2025-11-18T23:34:05Z, managed_by=tripleo_ansible, vendor=Red Hat, Inc., config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:03:32 localhost podman[105292]: unhealthy Nov 27 04:03:32 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:03:32 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:03:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:03:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:03:39 localhost podman[105354]: 2025-11-27 09:03:39.989257088 +0000 UTC m=+0.087766034 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, com.redhat.component=openstack-collectd-container, release=1761123044, description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, config_id=tripleo_step3, build-date=2025-11-18T22:51:28Z, container_name=collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}) Nov 27 04:03:39 localhost podman[105354]: 2025-11-27 09:03:39.998175668 +0000 UTC m=+0.096684604 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vendor=Red Hat, Inc., io.openshift.expose-services=, config_id=tripleo_step3, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, distribution-scope=public, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, version=17.1.12, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd, release=1761123044) Nov 27 04:03:40 localhost systemd[1]: tmp-crun.Fmf6PS.mount: Deactivated successfully. Nov 27 04:03:40 localhost podman[105355]: 2025-11-27 09:03:40.042912243 +0000 UTC m=+0.137899785 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, distribution-scope=public, managed_by=tripleo_ansible, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, name=rhosp17/openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, url=https://www.redhat.com, batch=17.1_20251118.1, version=17.1.12, vcs-type=git, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, build-date=2025-11-18T23:44:13Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid) Nov 27 04:03:40 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:03:40 localhost podman[105355]: 2025-11-27 09:03:40.081143563 +0000 UTC m=+0.176131105 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, io.openshift.expose-services=, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.component=openstack-iscsid-container, description=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, container_name=iscsid, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., url=https://www.redhat.com, name=rhosp17/openstack-iscsid, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, managed_by=tripleo_ansible, tcib_managed=true, config_id=tripleo_step3, vcs-type=git, release=1761123044) Nov 27 04:03:40 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:03:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:03:50 localhost podman[105393]: 2025-11-27 09:03:50.986160383 +0000 UTC m=+0.080745416 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.expose-services=, release=1761123044, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., container_name=metrics_qdr, name=rhosp17/openstack-qdrouterd, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, tcib_managed=true, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, distribution-scope=public, build-date=2025-11-18T22:49:46Z) Nov 27 04:03:51 localhost podman[105393]: 2025-11-27 09:03:51.203864625 +0000 UTC m=+0.298449688 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, version=17.1.12, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, maintainer=OpenStack TripleO Team, distribution-scope=public, vcs-type=git, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd) Nov 27 04:03:51 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:03:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:03:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:03:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:03:56 localhost podman[105422]: 2025-11-27 09:03:56.990014861 +0000 UTC m=+0.082926595 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, architecture=x86_64, release=1761123044, container_name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.component=openstack-ceilometer-ipmi-container, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc.) Nov 27 04:03:57 localhost podman[105422]: 2025-11-27 09:03:57.052945136 +0000 UTC m=+0.145856840 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, container_name=ceilometer_agent_ipmi, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, io.buildah.version=1.41.4, release=1761123044, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 04:03:57 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Deactivated successfully. Nov 27 04:03:57 localhost podman[105423]: 2025-11-27 09:03:57.1009927 +0000 UTC m=+0.188078287 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, distribution-scope=public, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, io.buildah.version=1.41.4, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, url=https://www.redhat.com, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, tcib_managed=true, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}) Nov 27 04:03:57 localhost podman[105424]: 2025-11-27 09:03:57.053752217 +0000 UTC m=+0.139418325 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_id=tripleo_step4, io.buildah.version=1.41.4, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.component=openstack-ceilometer-compute-container, vcs-type=git, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, url=https://www.redhat.com, architecture=x86_64, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, container_name=ceilometer_agent_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., name=rhosp17/openstack-ceilometer-compute) Nov 27 04:03:57 localhost podman[105423]: 2025-11-27 09:03:57.115090369 +0000 UTC m=+0.202175956 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, managed_by=tripleo_ansible, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, version=17.1.12, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=logrotate_crond, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, release=1761123044, url=https://www.redhat.com, vendor=Red Hat, Inc., name=rhosp17/openstack-cron) Nov 27 04:03:57 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:03:57 localhost podman[105424]: 2025-11-27 09:03:57.138063289 +0000 UTC m=+0.223729357 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, version=17.1.12, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, tcib_managed=true, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, release=1761123044, config_id=tripleo_step4) Nov 27 04:03:57 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:03:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:03:59 localhost podman[105492]: 2025-11-27 09:03:59.98380077 +0000 UTC m=+0.081661910 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, version=17.1.12, distribution-scope=public, tcib_managed=true, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, container_name=nova_migration_target, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, managed_by=tripleo_ansible, release=1761123044, build-date=2025-11-19T00:36:58Z) Nov 27 04:04:00 localhost podman[105492]: 2025-11-27 09:04:00.36399424 +0000 UTC m=+0.461855400 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, vcs-type=git, release=1761123044, com.redhat.component=openstack-nova-compute-container, version=17.1.12, url=https://www.redhat.com, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, config_id=tripleo_step4, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:04:00 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:04:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:04:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:04:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:04:03 localhost podman[105517]: 2025-11-27 09:04:03.004030234 +0000 UTC m=+0.091716941 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step5, container_name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, build-date=2025-11-19T00:36:58Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, name=rhosp17/openstack-nova-compute, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vcs-type=git, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, architecture=x86_64) Nov 27 04:04:03 localhost systemd[1]: tmp-crun.2fFEBI.mount: Deactivated successfully. Nov 27 04:04:03 localhost podman[105515]: 2025-11-27 09:04:03.063711211 +0000 UTC m=+0.153991118 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.expose-services=, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, version=17.1.12, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vendor=Red Hat, Inc., config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:04:03 localhost podman[105517]: 2025-11-27 09:04:03.080355909 +0000 UTC m=+0.168042616 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., release=1761123044, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, tcib_managed=true, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:04:03 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Deactivated successfully. Nov 27 04:04:03 localhost podman[105515]: 2025-11-27 09:04:03.105873407 +0000 UTC m=+0.196153344 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, url=https://www.redhat.com, version=17.1.12, architecture=x86_64, io.openshift.expose-services=, distribution-scope=public, release=1761123044, container_name=ovn_metadata_agent, io.buildah.version=1.41.4, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1) Nov 27 04:04:03 localhost podman[105515]: unhealthy Nov 27 04:04:03 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:03 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:04:03 localhost podman[105516]: 2025-11-27 09:04:03.153370056 +0000 UTC m=+0.243004096 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, tcib_managed=true, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, name=rhosp17/openstack-ovn-controller) Nov 27 04:04:03 localhost podman[105516]: 2025-11-27 09:04:03.171855083 +0000 UTC m=+0.261489103 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, version=17.1.12, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, container_name=ovn_controller, release=1761123044, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, vendor=Red Hat, Inc., distribution-scope=public, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:04:03 localhost podman[105516]: unhealthy Nov 27 04:04:03 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:03 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:04:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:04:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:04:10 localhost podman[105585]: 2025-11-27 09:04:10.987013237 +0000 UTC m=+0.085017261 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, name=rhosp17/openstack-collectd, config_id=tripleo_step3, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, release=1761123044, vcs-type=git, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, batch=17.1_20251118.1, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, container_name=collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:04:11 localhost podman[105585]: 2025-11-27 09:04:11.00088024 +0000 UTC m=+0.098884224 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, batch=17.1_20251118.1, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 collectd, architecture=x86_64, managed_by=tripleo_ansible, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, distribution-scope=public, com.redhat.component=openstack-collectd-container, release=1761123044, vcs-type=git, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, name=rhosp17/openstack-collectd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:04:11 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:04:11 localhost systemd[1]: tmp-crun.f0n09Z.mount: Deactivated successfully. Nov 27 04:04:11 localhost podman[105586]: 2025-11-27 09:04:11.107554963 +0000 UTC m=+0.198712873 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, description=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, distribution-scope=public, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.buildah.version=1.41.4, batch=17.1_20251118.1, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, com.redhat.component=openstack-iscsid-container, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, tcib_managed=true) Nov 27 04:04:11 localhost podman[105586]: 2025-11-27 09:04:11.120948694 +0000 UTC m=+0.212106614 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, maintainer=OpenStack TripleO Team, container_name=iscsid, batch=17.1_20251118.1, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, summary=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, vcs-type=git, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, config_id=tripleo_step3, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:04:11 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:04:13 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:04:13 localhost recover_tripleo_nova_virtqemud[105641]: 63639 Nov 27 04:04:13 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:04:13 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:04:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:04:22 localhost systemd[1]: tmp-crun.DcEDke.mount: Deactivated successfully. Nov 27 04:04:22 localhost podman[105704]: 2025-11-27 09:04:22.019993693 +0000 UTC m=+0.110375974 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, distribution-scope=public, build-date=2025-11-18T22:49:46Z, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, io.openshift.expose-services=, version=17.1.12, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, release=1761123044, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, vendor=Red Hat, Inc., config_id=tripleo_step1) Nov 27 04:04:22 localhost podman[105704]: 2025-11-27 09:04:22.258368784 +0000 UTC m=+0.348751025 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-qdrouterd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, release=1761123044, distribution-scope=public, io.openshift.expose-services=, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, batch=17.1_20251118.1) Nov 27 04:04:22 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:04:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:04:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:04:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:04:28 localhost podman[105734]: 2025-11-27 09:04:27.999195799 +0000 UTC m=+0.085782772 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.buildah.version=1.41.4, managed_by=tripleo_ansible, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, config_id=tripleo_step4, vcs-type=git, description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, name=rhosp17/openstack-cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:04:28 localhost podman[105734]: 2025-11-27 09:04:28.01482949 +0000 UTC m=+0.101416453 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, build-date=2025-11-18T22:49:32Z, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-cron-container, description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., architecture=x86_64, container_name=logrotate_crond, managed_by=tripleo_ansible, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-cron, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, io.openshift.expose-services=) Nov 27 04:04:28 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:04:28 localhost podman[105733]: 2025-11-27 09:04:28.058615129 +0000 UTC m=+0.144216425 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, distribution-scope=public, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, batch=17.1_20251118.1, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, url=https://www.redhat.com, release=1761123044, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 04:04:28 localhost podman[105733]: 2025-11-27 09:04:28.094934997 +0000 UTC m=+0.180536303 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, container_name=ceilometer_agent_ipmi, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, tcib_managed=true, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4) Nov 27 04:04:28 localhost podman[105733]: unhealthy Nov 27 04:04:28 localhost systemd[1]: tmp-crun.WluZvi.mount: Deactivated successfully. Nov 27 04:04:28 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:28 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:04:28 localhost podman[105735]: 2025-11-27 09:04:28.125327686 +0000 UTC m=+0.208937279 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, name=rhosp17/openstack-ceilometer-compute, release=1761123044, build-date=2025-11-19T00:11:48Z, tcib_managed=true, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, io.openshift.expose-services=, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, config_id=tripleo_step4, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, com.redhat.component=openstack-ceilometer-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1) Nov 27 04:04:28 localhost podman[105735]: 2025-11-27 09:04:28.160644197 +0000 UTC m=+0.244253800 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, batch=17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, io.buildah.version=1.41.4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, container_name=ceilometer_agent_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64) Nov 27 04:04:28 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:04:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:04:30 localhost podman[105804]: 2025-11-27 09:04:30.977890942 +0000 UTC m=+0.072629587 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, distribution-scope=public, container_name=nova_migration_target, build-date=2025-11-19T00:36:58Z, vcs-type=git, com.redhat.component=openstack-nova-compute-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, release=1761123044, tcib_managed=true, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 04:04:31 localhost podman[105804]: 2025-11-27 09:04:31.352031109 +0000 UTC m=+0.446769734 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., release=1761123044, tcib_managed=true, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, managed_by=tripleo_ansible, distribution-scope=public, architecture=x86_64, url=https://www.redhat.com, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:04:31 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:04:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:04:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:04:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:04:33 localhost systemd[1]: tmp-crun.HQH18h.mount: Deactivated successfully. Nov 27 04:04:34 localhost podman[105825]: 2025-11-27 09:04:34.00054832 +0000 UTC m=+0.094513026 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, release=1761123044, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, batch=17.1_20251118.1, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:04:34 localhost podman[105826]: 2025-11-27 09:04:34.043714543 +0000 UTC m=+0.134425981 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, vcs-type=git, build-date=2025-11-18T23:34:05Z, container_name=ovn_controller, io.buildah.version=1.41.4, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, architecture=x86_64, managed_by=tripleo_ansible) Nov 27 04:04:34 localhost podman[105825]: 2025-11-27 09:04:34.049270433 +0000 UTC m=+0.143235139 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, container_name=ovn_metadata_agent, io.openshift.expose-services=, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, version=17.1.12, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, release=1761123044, url=https://www.redhat.com, config_id=tripleo_step4) Nov 27 04:04:34 localhost podman[105825]: unhealthy Nov 27 04:04:34 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:34 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:04:34 localhost podman[105826]: 2025-11-27 09:04:34.089882336 +0000 UTC m=+0.180593834 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, release=1761123044, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, tcib_managed=true, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, com.redhat.component=openstack-ovn-controller-container) Nov 27 04:04:34 localhost podman[105826]: unhealthy Nov 27 04:04:34 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:34 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:04:34 localhost podman[105827]: 2025-11-27 09:04:34.142933585 +0000 UTC m=+0.231290120 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, batch=17.1_20251118.1, vendor=Red Hat, Inc., url=https://www.redhat.com, distribution-scope=public, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, io.openshift.expose-services=, release=1761123044, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container) Nov 27 04:04:34 localhost podman[105827]: 2025-11-27 09:04:34.190858936 +0000 UTC m=+0.279215451 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, release=1761123044, url=https://www.redhat.com, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-compute, config_id=tripleo_step5, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.expose-services=, vendor=Red Hat, Inc., version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:04:34 localhost podman[105827]: unhealthy Nov 27 04:04:34 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:34 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:04:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:04:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:04:41 localhost podman[105888]: 2025-11-27 09:04:41.988183529 +0000 UTC m=+0.084320193 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, tcib_managed=true, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, com.redhat.component=openstack-collectd-container, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_id=tripleo_step3, vcs-type=git, container_name=collectd, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, url=https://www.redhat.com, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-collectd) Nov 27 04:04:41 localhost podman[105888]: 2025-11-27 09:04:41.997555831 +0000 UTC m=+0.093692495 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, release=1761123044, version=17.1.12, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.buildah.version=1.41.4, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, name=rhosp17/openstack-collectd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, tcib_managed=true, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd) Nov 27 04:04:42 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:04:42 localhost systemd[1]: tmp-crun.2iiRmy.mount: Deactivated successfully. Nov 27 04:04:42 localhost podman[105889]: 2025-11-27 09:04:42.052070999 +0000 UTC m=+0.145169640 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, com.redhat.component=openstack-iscsid-container, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, architecture=x86_64, container_name=iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid) Nov 27 04:04:42 localhost podman[105889]: 2025-11-27 09:04:42.087619586 +0000 UTC m=+0.180718177 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.buildah.version=1.41.4, com.redhat.component=openstack-iscsid-container, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=iscsid, build-date=2025-11-18T23:44:13Z, release=1761123044, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, distribution-scope=public, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vcs-type=git, config_id=tripleo_step3) Nov 27 04:04:42 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:04:44 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60574 DF PROTO=TCP SPT=52432 DPT=9100 SEQ=1306393959 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12792AF00000000001030307) Nov 27 04:04:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19931 DF PROTO=TCP SPT=56996 DPT=9102 SEQ=71622133 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12792EC80000000001030307) Nov 27 04:04:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60575 DF PROTO=TCP SPT=52432 DPT=9100 SEQ=1306393959 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12792EDA0000000001030307) Nov 27 04:04:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19932 DF PROTO=TCP SPT=56996 DPT=9102 SEQ=71622133 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127932DA0000000001030307) Nov 27 04:04:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60576 DF PROTO=TCP SPT=52432 DPT=9100 SEQ=1306393959 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127936DB0000000001030307) Nov 27 04:04:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19933 DF PROTO=TCP SPT=56996 DPT=9102 SEQ=71622133 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12793ADA0000000001030307) Nov 27 04:04:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60577 DF PROTO=TCP SPT=52432 DPT=9100 SEQ=1306393959 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279469B0000000001030307) Nov 27 04:04:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44772 DF PROTO=TCP SPT=56752 DPT=9101 SEQ=3192826427 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12794A920000000001030307) Nov 27 04:04:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19934 DF PROTO=TCP SPT=56996 DPT=9102 SEQ=71622133 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12794A9A0000000001030307) Nov 27 04:04:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:04:52 localhost podman[105926]: 2025-11-27 09:04:52.982932635 +0000 UTC m=+0.082800930 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, vendor=Red Hat, Inc., architecture=x86_64, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-qdrouterd-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, config_id=tripleo_step1, container_name=metrics_qdr, maintainer=OpenStack TripleO Team, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, name=rhosp17/openstack-qdrouterd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=) Nov 27 04:04:53 localhost podman[105926]: 2025-11-27 09:04:53.178830901 +0000 UTC m=+0.278699166 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, tcib_managed=true, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, version=17.1.12, name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, batch=17.1_20251118.1, vcs-type=git, description=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, distribution-scope=public, container_name=metrics_qdr, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, com.redhat.component=openstack-qdrouterd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, build-date=2025-11-18T22:49:46Z) Nov 27 04:04:53 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:04:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44773 DF PROTO=TCP SPT=56752 DPT=9101 SEQ=3192826427 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12794E9A0000000001030307) Nov 27 04:04:55 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44774 DF PROTO=TCP SPT=56752 DPT=9101 SEQ=3192826427 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279569A0000000001030307) Nov 27 04:04:55 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44986 DF PROTO=TCP SPT=44614 DPT=9105 SEQ=2127559193 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127956B30000000001030307) Nov 27 04:04:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44987 DF PROTO=TCP SPT=44614 DPT=9105 SEQ=2127559193 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12795ADB0000000001030307) Nov 27 04:04:58 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44988 DF PROTO=TCP SPT=44614 DPT=9105 SEQ=2127559193 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127962DA0000000001030307) Nov 27 04:04:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:04:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:04:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:04:58 localhost podman[105955]: 2025-11-27 09:04:58.9965862 +0000 UTC m=+0.086606424 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-ipmi-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, version=17.1.12, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, vcs-type=git, config_id=tripleo_step4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-ceilometer-ipmi) Nov 27 04:04:59 localhost podman[105955]: 2025-11-27 09:04:59.052770433 +0000 UTC m=+0.142790617 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12, container_name=ceilometer_agent_ipmi, architecture=x86_64, tcib_managed=true, url=https://www.redhat.com, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-ipmi-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, build-date=2025-11-19T00:12:45Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 04:04:59 localhost systemd[1]: tmp-crun.NVkpLa.mount: Deactivated successfully. Nov 27 04:04:59 localhost podman[105955]: unhealthy Nov 27 04:04:59 localhost podman[105957]: 2025-11-27 09:04:59.065559388 +0000 UTC m=+0.147399402 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, distribution-scope=public, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, io.openshift.expose-services=, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, managed_by=tripleo_ansible, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, maintainer=OpenStack TripleO Team, tcib_managed=true, release=1761123044) Nov 27 04:04:59 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:04:59 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:04:59 localhost podman[105956]: 2025-11-27 09:04:59.101737602 +0000 UTC m=+0.187706126 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_id=tripleo_step4, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, name=rhosp17/openstack-cron, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, container_name=logrotate_crond, distribution-scope=public, managed_by=tripleo_ansible) Nov 27 04:04:59 localhost podman[105956]: 2025-11-27 09:04:59.111104954 +0000 UTC m=+0.197073548 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vendor=Red Hat, Inc., distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.buildah.version=1.41.4, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, architecture=x86_64, version=17.1.12, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:32Z, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, name=rhosp17/openstack-cron, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:04:59 localhost podman[105957]: 2025-11-27 09:04:59.119777868 +0000 UTC m=+0.201617892 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, vcs-type=git, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, build-date=2025-11-19T00:11:48Z, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, distribution-scope=public, managed_by=tripleo_ansible, container_name=ceilometer_agent_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, url=https://www.redhat.com, name=rhosp17/openstack-ceilometer-compute) Nov 27 04:04:59 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:04:59 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:04:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44775 DF PROTO=TCP SPT=56752 DPT=9101 SEQ=3192826427 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279665A0000000001030307) Nov 27 04:05:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60578 DF PROTO=TCP SPT=52432 DPT=9100 SEQ=1306393959 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127967DB0000000001030307) Nov 27 04:05:01 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19935 DF PROTO=TCP SPT=56996 DPT=9102 SEQ=71622133 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12796BDA0000000001030307) Nov 27 04:05:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:05:01 localhost podman[106024]: 2025-11-27 09:05:01.995067466 +0000 UTC m=+0.088626957 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, container_name=nova_migration_target, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, release=1761123044, name=rhosp17/openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.openshift.expose-services=, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}) Nov 27 04:05:02 localhost podman[106024]: 2025-11-27 09:05:02.422995642 +0000 UTC m=+0.516555143 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, tcib_managed=true, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, vcs-type=git) Nov 27 04:05:02 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:05:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44989 DF PROTO=TCP SPT=44614 DPT=9105 SEQ=2127559193 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279729A0000000001030307) Nov 27 04:05:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:05:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:05:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:05:05 localhost systemd[1]: tmp-crun.O1dsge.mount: Deactivated successfully. Nov 27 04:05:05 localhost podman[106049]: 2025-11-27 09:05:05.011470556 +0000 UTC m=+0.087615110 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, com.redhat.component=openstack-nova-compute-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, release=1761123044, version=17.1.12, distribution-scope=public, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, vcs-type=git) Nov 27 04:05:05 localhost podman[106048]: 2025-11-27 09:05:05.057217628 +0000 UTC m=+0.135290935 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, container_name=ovn_controller, config_id=tripleo_step4, vendor=Red Hat, Inc., org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, vcs-type=git, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:05:05 localhost podman[106048]: 2025-11-27 09:05:05.098942112 +0000 UTC m=+0.177015379 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_id=tripleo_step4, vendor=Red Hat, Inc., batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, build-date=2025-11-18T23:34:05Z, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, release=1761123044, version=17.1.12, io.openshift.expose-services=, maintainer=OpenStack TripleO Team) Nov 27 04:05:05 localhost podman[106048]: unhealthy Nov 27 04:05:05 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:05 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:05:05 localhost podman[106047]: 2025-11-27 09:05:05.114638645 +0000 UTC m=+0.197274104 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.buildah.version=1.41.4, io.openshift.expose-services=, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, batch=17.1_20251118.1, vendor=Red Hat, Inc., container_name=ovn_metadata_agent, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, distribution-scope=public, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 04:05:05 localhost podman[106049]: 2025-11-27 09:05:05.138221599 +0000 UTC m=+0.214366153 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, release=1761123044, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, version=17.1.12, batch=17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, distribution-scope=public, container_name=nova_compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, io.buildah.version=1.41.4, vcs-type=git, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:05:05 localhost podman[106049]: unhealthy Nov 27 04:05:05 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:05 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:05:05 localhost podman[106047]: 2025-11-27 09:05:05.162895114 +0000 UTC m=+0.245530543 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, batch=17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, io.buildah.version=1.41.4, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, tcib_managed=true, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git) Nov 27 04:05:05 localhost podman[106047]: unhealthy Nov 27 04:05:05 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:05 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:05:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44776 DF PROTO=TCP SPT=56752 DPT=9101 SEQ=3192826427 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127985DA0000000001030307) Nov 27 04:05:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6292 DF PROTO=TCP SPT=46314 DPT=9882 SEQ=2552276907 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127989660000000001030307) Nov 27 04:05:09 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6293 DF PROTO=TCP SPT=46314 DPT=9882 SEQ=2552276907 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12798D5A0000000001030307) Nov 27 04:05:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44990 DF PROTO=TCP SPT=44614 DPT=9105 SEQ=2127559193 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127993DB0000000001030307) Nov 27 04:05:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6294 DF PROTO=TCP SPT=46314 DPT=9882 SEQ=2552276907 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279955A0000000001030307) Nov 27 04:05:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:05:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:05:12 localhost podman[106107]: 2025-11-27 09:05:12.994913731 +0000 UTC m=+0.089145532 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.openshift.expose-services=, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:51:28Z, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, container_name=collectd, batch=17.1_20251118.1, release=1761123044, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd) Nov 27 04:05:13 localhost podman[106107]: 2025-11-27 09:05:13.033051698 +0000 UTC m=+0.127283449 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, description=Red Hat OpenStack Platform 17.1 collectd, vendor=Red Hat, Inc., release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, io.openshift.expose-services=, architecture=x86_64, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, batch=17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, name=rhosp17/openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, build-date=2025-11-18T22:51:28Z, config_id=tripleo_step3) Nov 27 04:05:13 localhost podman[106108]: 2025-11-27 09:05:13.044192338 +0000 UTC m=+0.135059059 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, container_name=iscsid, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, managed_by=tripleo_ansible, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, version=17.1.12, tcib_managed=true, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, maintainer=OpenStack TripleO Team, vcs-type=git, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.buildah.version=1.41.4, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc.) Nov 27 04:05:13 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:05:13 localhost podman[106108]: 2025-11-27 09:05:13.058900654 +0000 UTC m=+0.149767325 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, vendor=Red Hat, Inc., version=17.1.12, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, vcs-type=git, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, summary=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, build-date=2025-11-18T23:44:13Z, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, architecture=x86_64, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, batch=17.1_20251118.1, release=1761123044, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 04:05:13 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:05:14 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44998 DF PROTO=TCP SPT=46848 DPT=9100 SEQ=2925613904 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279A0200000000001030307) Nov 27 04:05:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47252 DF PROTO=TCP SPT=56916 DPT=9102 SEQ=775630294 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279A3F90000000001030307) Nov 27 04:05:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45000 DF PROTO=TCP SPT=46848 DPT=9100 SEQ=2925613904 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279AC1A0000000001030307) Nov 27 04:05:21 localhost sshd[106224]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:05:21 localhost systemd-logind[761]: New session 37 of user zuul. Nov 27 04:05:21 localhost systemd[1]: Started Session 37 of User zuul. Nov 27 04:05:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45001 DF PROTO=TCP SPT=46848 DPT=9100 SEQ=2925613904 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279BBDA0000000001030307) Nov 27 04:05:22 localhost python3.9[106319]: ansible-ansible.builtin.stat Invoked with path=/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova/nova.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:05:22 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:05:22 localhost recover_tripleo_nova_virtqemud[106415]: 63639 Nov 27 04:05:22 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:05:22 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:05:22 localhost python3.9[106413]: ansible-ansible.legacy.command Invoked with cmd=python3 -c "import configparser as c; p = c.ConfigParser(strict=False); p.read('/var/lib/config-data/puppet-generated/nova_libvirt/etc/nova/nova.conf'); print(p['DEFAULT']['host'])"#012 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:05:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:05:23 localhost systemd[1]: tmp-crun.8r4nU3.mount: Deactivated successfully. Nov 27 04:05:23 localhost podman[106508]: 2025-11-27 09:05:23.593013895 +0000 UTC m=+0.103951361 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, name=rhosp17/openstack-qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=tripleo_step1, io.openshift.expose-services=, url=https://www.redhat.com, architecture=x86_64, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, version=17.1.12, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, container_name=metrics_qdr) Nov 27 04:05:23 localhost python3.9[106509]: ansible-ansible.builtin.stat Invoked with path=/var/lib/config-data/puppet-generated/neutron/etc/neutron/neutron.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:05:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17677 DF PROTO=TCP SPT=40194 DPT=9101 SEQ=1200696709 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279C3DA0000000001030307) Nov 27 04:05:23 localhost podman[106508]: 2025-11-27 09:05:23.804810859 +0000 UTC m=+0.315748265 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, maintainer=OpenStack TripleO Team, config_id=tripleo_step1, batch=17.1_20251118.1, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, version=17.1.12, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, tcib_managed=true, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-qdrouterd, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vcs-type=git, architecture=x86_64, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:05:23 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:05:24 localhost python3.9[106630]: ansible-ansible.legacy.command Invoked with cmd=python3 -c "import configparser as c; p = c.ConfigParser(strict=False); p.read('/var/lib/config-data/puppet-generated/neutron/etc/neutron/neutron.conf'); print(p['DEFAULT']['host'])"#012 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:05:25 localhost python3.9[106724]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/hostname -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:05:26 localhost python3.9[106815]: ansible-ansible.builtin.slurp Invoked with src=/proc/cmdline Nov 27 04:05:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36679 DF PROTO=TCP SPT=47178 DPT=9105 SEQ=2662127450 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279CFDB0000000001030307) Nov 27 04:05:28 localhost python3.9[106905]: ansible-ansible.builtin.stat Invoked with path=/etc/tuned/active_profile follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:05:28 localhost python3.9[106997]: ansible-ansible.builtin.slurp Invoked with src=/etc/tuned/active_profile Nov 27 04:05:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17679 DF PROTO=TCP SPT=40194 DPT=9101 SEQ=1200696709 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279DB9A0000000001030307) Nov 27 04:05:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:05:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:05:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:05:29 localhost python3.9[107087]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:05:29 localhost systemd[1]: tmp-crun.x2BPrf.mount: Deactivated successfully. Nov 27 04:05:30 localhost podman[107089]: 2025-11-27 09:05:29.999384716 +0000 UTC m=+0.076455860 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, vcs-type=git, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, version=17.1.12, name=rhosp17/openstack-cron, vendor=Red Hat, Inc., vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:05:30 localhost podman[107089]: 2025-11-27 09:05:30.013787003 +0000 UTC m=+0.090858137 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:32Z, io.openshift.expose-services=, io.buildah.version=1.41.4, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, version=17.1.12, url=https://www.redhat.com) Nov 27 04:05:30 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:05:30 localhost systemd[1]: tmp-crun.m1FzU8.mount: Deactivated successfully. Nov 27 04:05:30 localhost podman[107088]: 2025-11-27 09:05:30.061492268 +0000 UTC m=+0.139445657 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=healthy, version=17.1.12, build-date=2025-11-19T00:12:45Z, name=rhosp17/openstack-ceilometer-ipmi, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, vcs-type=git, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, container_name=ceilometer_agent_ipmi, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, distribution-scope=public, config_id=tripleo_step4, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 04:05:30 localhost podman[107088]: 2025-11-27 09:05:30.091939788 +0000 UTC m=+0.169893277 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.buildah.version=1.41.4, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, version=17.1.12, com.redhat.component=openstack-ceilometer-ipmi-container, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, architecture=x86_64) Nov 27 04:05:30 localhost podman[107088]: unhealthy Nov 27 04:05:30 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:30 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:05:30 localhost podman[107090]: 2025-11-27 09:05:30.113816208 +0000 UTC m=+0.187643735 container health_status f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, vendor=Red Hat, Inc., io.buildah.version=1.41.4, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, name=rhosp17/openstack-ceilometer-compute, distribution-scope=public, release=1761123044, architecture=x86_64, url=https://www.redhat.com, com.redhat.component=openstack-ceilometer-compute-container, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:11:48Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1) Nov 27 04:05:30 localhost podman[107090]: 2025-11-27 09:05:30.147014072 +0000 UTC m=+0.220841609 container exec_died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, io.buildah.version=1.41.4, container_name=ceilometer_agent_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:11:48Z, config_id=tripleo_step4, managed_by=tripleo_ansible, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, name=rhosp17/openstack-ceilometer-compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, url=https://www.redhat.com, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, distribution-scope=public, tcib_managed=true, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:05:30 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Deactivated successfully. Nov 27 04:05:30 localhost python3.9[107207]: ansible-ansible.legacy.dnf Invoked with name=['systemd-container'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:05:31 localhost systemd[1]: session-37.scope: Deactivated successfully. Nov 27 04:05:31 localhost systemd[1]: session-37.scope: Consumed 4.833s CPU time. Nov 27 04:05:31 localhost systemd-logind[761]: Session 37 logged out. Waiting for processes to exit. Nov 27 04:05:31 localhost systemd-logind[761]: Removed session 37. Nov 27 04:05:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36681 DF PROTO=TCP SPT=47178 DPT=9105 SEQ=2662127450 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279E79A0000000001030307) Nov 27 04:05:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:05:32 localhost podman[107223]: 2025-11-27 09:05:32.98070777 +0000 UTC m=+0.082694998 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, distribution-scope=public, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, name=rhosp17/openstack-nova-compute, managed_by=tripleo_ansible) Nov 27 04:05:33 localhost podman[107223]: 2025-11-27 09:05:33.347752506 +0000 UTC m=+0.449739754 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, distribution-scope=public, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_migration_target, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_id=tripleo_step4, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, release=1761123044, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64) Nov 27 04:05:33 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:05:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:05:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:05:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:05:36 localhost podman[107249]: 2025-11-27 09:05:35.999619158 +0000 UTC m=+0.084963069 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=healthy, vcs-type=git, com.redhat.component=openstack-nova-compute-container, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, name=rhosp17/openstack-nova-compute, io.openshift.expose-services=, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, release=1761123044, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, architecture=x86_64, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, version=17.1.12, vendor=Red Hat, Inc., config_id=tripleo_step5, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:05:36 localhost podman[107249]: 2025-11-27 09:05:36.015890656 +0000 UTC m=+0.101234587 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, distribution-scope=public, tcib_managed=true, url=https://www.redhat.com, version=17.1.12, name=rhosp17/openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container) Nov 27 04:05:36 localhost podman[107249]: unhealthy Nov 27 04:05:36 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:36 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:05:36 localhost systemd[1]: tmp-crun.lnW7Q3.mount: Deactivated successfully. Nov 27 04:05:36 localhost podman[107248]: 2025-11-27 09:05:36.143699378 +0000 UTC m=+0.228495135 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, io.buildah.version=1.41.4, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=ovn_controller, io.openshift.expose-services=, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, vcs-type=git, version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., architecture=x86_64, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:05:36 localhost podman[107247]: 2025-11-27 09:05:36.104932603 +0000 UTC m=+0.192751381 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4, architecture=x86_64, batch=17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, tcib_managed=true, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, managed_by=tripleo_ansible, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:05:36 localhost podman[107248]: 2025-11-27 09:05:36.158935189 +0000 UTC m=+0.243730966 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, name=rhosp17/openstack-ovn-controller, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_id=tripleo_step4, io.openshift.expose-services=, tcib_managed=true, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T23:34:05Z, distribution-scope=public, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, container_name=ovn_controller, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:05:36 localhost podman[107248]: unhealthy Nov 27 04:05:36 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:36 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:05:36 localhost podman[107247]: 2025-11-27 09:05:36.188063663 +0000 UTC m=+0.275882441 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, name=rhosp17/openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, version=17.1.12, distribution-scope=public, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., release=1761123044, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:05:36 localhost podman[107247]: unhealthy Nov 27 04:05:36 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:05:36 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:05:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17680 DF PROTO=TCP SPT=40194 DPT=9101 SEQ=1200696709 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279FBDA0000000001030307) Nov 27 04:05:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6089 DF PROTO=TCP SPT=44720 DPT=9882 SEQ=1124129469 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1279FE970000000001030307) Nov 27 04:05:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6091 DF PROTO=TCP SPT=44720 DPT=9882 SEQ=1124129469 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A0A9A0000000001030307) Nov 27 04:05:43 localhost sshd[107312]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:05:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:05:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:05:43 localhost systemd-logind[761]: New session 38 of user zuul. Nov 27 04:05:43 localhost systemd[1]: Started Session 38 of User zuul. Nov 27 04:05:43 localhost systemd[1]: tmp-crun.r54AXP.mount: Deactivated successfully. Nov 27 04:05:43 localhost podman[107314]: 2025-11-27 09:05:43.361949494 +0000 UTC m=+0.109103149 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, batch=17.1_20251118.1, io.openshift.expose-services=, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, name=rhosp17/openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, version=17.1.12, vendor=Red Hat, Inc., tcib_managed=true, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, container_name=collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3) Nov 27 04:05:43 localhost podman[107314]: 2025-11-27 09:05:43.372436897 +0000 UTC m=+0.119590562 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, managed_by=tripleo_ansible, release=1761123044, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, name=rhosp17/openstack-collectd, description=Red Hat OpenStack Platform 17.1 collectd, maintainer=OpenStack TripleO Team, build-date=2025-11-18T22:51:28Z, version=17.1.12, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, url=https://www.redhat.com, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, architecture=x86_64) Nov 27 04:05:43 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:05:43 localhost systemd[1]: tmp-crun.1SOdGT.mount: Deactivated successfully. Nov 27 04:05:43 localhost podman[107315]: 2025-11-27 09:05:43.464916037 +0000 UTC m=+0.197238693 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, vendor=Red Hat, Inc., io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, com.redhat.component=openstack-iscsid-container, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, description=Red Hat OpenStack Platform 17.1 iscsid, build-date=2025-11-18T23:44:13Z, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, container_name=iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, version=17.1.12) Nov 27 04:05:43 localhost podman[107315]: 2025-11-27 09:05:43.505031168 +0000 UTC m=+0.237353854 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, architecture=x86_64, build-date=2025-11-18T23:44:13Z, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, container_name=iscsid, io.buildah.version=1.41.4, tcib_managed=true, vendor=Red Hat, Inc., vcs-type=git, summary=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, version=17.1.12, com.redhat.component=openstack-iscsid-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:05:43 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:05:44 localhost python3.9[107445]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:05:44 localhost systemd[1]: Reloading. Nov 27 04:05:44 localhost systemd-rc-local-generator[107472]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:05:44 localhost systemd-sysv-generator[107476]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:05:44 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:05:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62237 DF PROTO=TCP SPT=54964 DPT=9102 SEQ=4176452053 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A19290000000001030307) Nov 27 04:05:45 localhost python3.9[107571]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:05:45 localhost network[107588]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:05:45 localhost network[107589]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:05:45 localhost network[107590]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:05:46 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:05:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46077 DF PROTO=TCP SPT=60732 DPT=9100 SEQ=2972935877 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A215A0000000001030307) Nov 27 04:05:49 localhost python3.9[107787]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:05:49 localhost network[107804]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:05:49 localhost network[107805]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:05:49 localhost network[107806]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:05:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46078 DF PROTO=TCP SPT=60732 DPT=9100 SEQ=2972935877 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A311B0000000001030307) Nov 27 04:05:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:05:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9243 DF PROTO=TCP SPT=48390 DPT=9101 SEQ=2241878083 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A391A0000000001030307) Nov 27 04:05:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:05:54 localhost podman[107941]: 2025-11-27 09:05:54.014053312 +0000 UTC m=+0.100694653 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, architecture=x86_64, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, version=17.1.12, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., io.openshift.expose-services=, name=rhosp17/openstack-qdrouterd, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, url=https://www.redhat.com, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.component=openstack-qdrouterd-container, io.buildah.version=1.41.4) Nov 27 04:05:54 localhost podman[107941]: 2025-11-27 09:05:54.208006577 +0000 UTC m=+0.294647958 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., config_id=tripleo_step1, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, build-date=2025-11-18T22:49:46Z, vcs-type=git, managed_by=tripleo_ansible, release=1761123044, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, com.redhat.component=openstack-qdrouterd-container, distribution-scope=public, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 04:05:54 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:05:54 localhost python3.9[108032]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_ceilometer_agent_compute.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:05:54 localhost systemd[1]: Reloading. Nov 27 04:05:54 localhost systemd-sysv-generator[108061]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:05:54 localhost systemd-rc-local-generator[108056]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:05:54 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:05:55 localhost systemd[1]: Stopping ceilometer_agent_compute container... Nov 27 04:05:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11693 DF PROTO=TCP SPT=44656 DPT=9105 SEQ=1903378834 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A451B0000000001030307) Nov 27 04:05:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9245 DF PROTO=TCP SPT=48390 DPT=9101 SEQ=2241878083 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A50DA0000000001030307) Nov 27 04:06:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:06:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:06:00 localhost systemd[1]: tmp-crun.HC47No.mount: Deactivated successfully. Nov 27 04:06:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:06:00 localhost podman[108088]: 2025-11-27 09:06:00.250692302 +0000 UTC m=+0.097270221 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=unhealthy, vendor=Red Hat, Inc., release=1761123044, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, architecture=x86_64, name=rhosp17/openstack-ceilometer-ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-type=git, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:12:45Z, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, config_id=tripleo_step4, distribution-scope=public, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-ipmi-container) Nov 27 04:06:00 localhost podman[108115]: Error: container f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a is not running Nov 27 04:06:00 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Main process exited, code=exited, status=125/n/a Nov 27 04:06:00 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed with result 'exit-code'. Nov 27 04:06:00 localhost systemd[1]: tmp-crun.iW7UDt.mount: Deactivated successfully. Nov 27 04:06:00 localhost podman[108089]: 2025-11-27 09:06:00.404507595 +0000 UTC m=+0.246132701 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, name=rhosp17/openstack-cron, com.redhat.component=openstack-cron-container, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, container_name=logrotate_crond, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, tcib_managed=true, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, build-date=2025-11-18T22:49:32Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:06:00 localhost podman[108088]: 2025-11-27 09:06:00.413216689 +0000 UTC m=+0.259794658 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, tcib_managed=true, com.redhat.component=openstack-ceilometer-ipmi-container, release=1761123044, build-date=2025-11-19T00:12:45Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, distribution-scope=public, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, url=https://www.redhat.com, container_name=ceilometer_agent_ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-ceilometer-ipmi, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}) Nov 27 04:06:00 localhost podman[108088]: unhealthy Nov 27 04:06:00 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:00 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:06:00 localhost podman[108089]: 2025-11-27 09:06:00.468788736 +0000 UTC m=+0.310413842 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, version=17.1.12, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, batch=17.1_20251118.1, com.redhat.component=openstack-cron-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=logrotate_crond, release=1761123044, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z) Nov 27 04:06:00 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:06:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11695 DF PROTO=TCP SPT=44656 DPT=9105 SEQ=1903378834 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A5CDA0000000001030307) Nov 27 04:06:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:06:03 localhost podman[108148]: 2025-11-27 09:06:03.478242268 +0000 UTC m=+0.074278322 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, architecture=x86_64, distribution-scope=public, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, url=https://www.redhat.com, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, managed_by=tripleo_ansible) Nov 27 04:06:03 localhost podman[108148]: 2025-11-27 09:06:03.801896934 +0000 UTC m=+0.397932978 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, distribution-scope=public, version=17.1.12, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-compute, release=1761123044, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 04:06:03 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:06:06 localhost systemd[1]: tmp-crun.Yc6Iqy.mount: Deactivated successfully. Nov 27 04:06:06 localhost podman[108171]: 2025-11-27 09:06:06.246482174 +0000 UTC m=+0.088054683 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, architecture=x86_64, container_name=nova_compute, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, config_id=tripleo_step5, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, version=17.1.12) Nov 27 04:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:06:06 localhost podman[108171]: 2025-11-27 09:06:06.29798828 +0000 UTC m=+0.139560759 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, version=17.1.12, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, batch=17.1_20251118.1, io.buildah.version=1.41.4, com.redhat.component=openstack-nova-compute-container, container_name=nova_compute, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, config_id=tripleo_step5, tcib_managed=true) Nov 27 04:06:06 localhost podman[108171]: unhealthy Nov 27 04:06:06 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:06 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:06:06 localhost systemd[1]: tmp-crun.ak5vU0.mount: Deactivated successfully. Nov 27 04:06:06 localhost podman[108194]: 2025-11-27 09:06:06.384606383 +0000 UTC m=+0.114304089 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, architecture=x86_64, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vendor=Red Hat, Inc., release=1761123044, vcs-type=git, tcib_managed=true, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 04:06:06 localhost podman[108194]: 2025-11-27 09:06:06.423386638 +0000 UTC m=+0.153084334 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, batch=17.1_20251118.1, name=rhosp17/openstack-ovn-controller, distribution-scope=public, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vendor=Red Hat, Inc., release=1761123044, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, vcs-type=git, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, url=https://www.redhat.com, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 04:06:06 localhost podman[108194]: unhealthy Nov 27 04:06:06 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:06 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:06:06 localhost podman[108193]: 2025-11-27 09:06:06.429228875 +0000 UTC m=+0.161010397 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, container_name=ovn_metadata_agent, version=17.1.12, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, batch=17.1_20251118.1, vcs-type=git, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc.) Nov 27 04:06:06 localhost podman[108193]: 2025-11-27 09:06:06.509142848 +0000 UTC m=+0.240924360 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, tcib_managed=true, version=17.1.12, config_id=tripleo_step4, vcs-type=git, container_name=ovn_metadata_agent, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64) Nov 27 04:06:06 localhost podman[108193]: unhealthy Nov 27 04:06:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:06:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9246 DF PROTO=TCP SPT=48390 DPT=9101 SEQ=2241878083 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A71DA0000000001030307) Nov 27 04:06:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18328 DF PROTO=TCP SPT=35062 DPT=9882 SEQ=2901892248 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A73C70000000001030307) Nov 27 04:06:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18330 DF PROTO=TCP SPT=35062 DPT=9882 SEQ=2901892248 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A7FDA0000000001030307) Nov 27 04:06:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:06:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:06:13 localhost systemd[1]: tmp-crun.SvNV2e.mount: Deactivated successfully. Nov 27 04:06:13 localhost podman[108233]: 2025-11-27 09:06:13.998083104 +0000 UTC m=+0.090977691 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, description=Red Hat OpenStack Platform 17.1 collectd, summary=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, container_name=collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, build-date=2025-11-18T22:51:28Z, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-collectd-container, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.openshift.expose-services=, io.buildah.version=1.41.4, distribution-scope=public, config_id=tripleo_step3, vendor=Red Hat, Inc., managed_by=tripleo_ansible) Nov 27 04:06:14 localhost podman[108234]: 2025-11-27 09:06:14.047601258 +0000 UTC m=+0.136803786 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, architecture=x86_64, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, distribution-scope=public, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_id=tripleo_step3, container_name=iscsid, vcs-type=git, tcib_managed=true, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, release=1761123044, name=rhosp17/openstack-iscsid, version=17.1.12, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:06:14 localhost podman[108234]: 2025-11-27 09:06:14.056762384 +0000 UTC m=+0.145964882 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, architecture=x86_64, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, container_name=iscsid, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, tcib_managed=true, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, vendor=Red Hat, Inc., managed_by=tripleo_ansible) Nov 27 04:06:14 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:06:14 localhost podman[108233]: 2025-11-27 09:06:14.113638546 +0000 UTC m=+0.206533133 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, vcs-type=git, description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, io.openshift.expose-services=, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, tcib_managed=true, container_name=collectd, batch=17.1_20251118.1, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:06:14 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:06:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42543 DF PROTO=TCP SPT=34628 DPT=9102 SEQ=1543528945 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A8E590000000001030307) Nov 27 04:06:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11036 DF PROTO=TCP SPT=41322 DPT=9100 SEQ=2675897657 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127A969B0000000001030307) Nov 27 04:06:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11037 DF PROTO=TCP SPT=41322 DPT=9100 SEQ=2675897657 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AA65A0000000001030307) Nov 27 04:06:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47881 DF PROTO=TCP SPT=55916 DPT=9101 SEQ=2369423777 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AAE1B0000000001030307) Nov 27 04:06:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:06:24 localhost podman[108350]: 2025-11-27 09:06:24.497789749 +0000 UTC m=+0.086047089 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, config_id=tripleo_step1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, io.openshift.expose-services=, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-qdrouterd, release=1761123044, version=17.1.12, batch=17.1_20251118.1, url=https://www.redhat.com, vendor=Red Hat, Inc., container_name=metrics_qdr, io.buildah.version=1.41.4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64) Nov 27 04:06:24 localhost podman[108350]: 2025-11-27 09:06:24.692956906 +0000 UTC m=+0.281214236 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, container_name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vcs-type=git, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., name=rhosp17/openstack-qdrouterd, architecture=x86_64, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, tcib_managed=true, com.redhat.component=openstack-qdrouterd-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd) Nov 27 04:06:24 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:06:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17682 DF PROTO=TCP SPT=40194 DPT=9101 SEQ=1200696709 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AB9DB0000000001030307) Nov 27 04:06:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11038 DF PROTO=TCP SPT=41322 DPT=9100 SEQ=2675897657 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AC5DA0000000001030307) Nov 27 04:06:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:06:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:06:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:06:30 localhost podman[108381]: Error: container f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a is not running Nov 27 04:06:30 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Main process exited, code=exited, status=125/n/a Nov 27 04:06:30 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed with result 'exit-code'. Nov 27 04:06:30 localhost systemd[1]: tmp-crun.g52sak.mount: Deactivated successfully. Nov 27 04:06:30 localhost systemd[1]: tmp-crun.0IHg9P.mount: Deactivated successfully. Nov 27 04:06:30 localhost podman[108379]: 2025-11-27 09:06:30.788282178 +0000 UTC m=+0.133700672 container health_status ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, health_status=unhealthy, container_name=ceilometer_agent_ipmi, config_id=tripleo_step4, vcs-type=git, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ceilometer-ipmi, build-date=2025-11-19T00:12:45Z, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, version=17.1.12, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-ipmi-container, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi) Nov 27 04:06:30 localhost podman[108380]: 2025-11-27 09:06:30.760121759 +0000 UTC m=+0.101714370 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, container_name=logrotate_crond, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, config_id=tripleo_step4, build-date=2025-11-18T22:49:32Z, name=rhosp17/openstack-cron, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:06:30 localhost podman[108379]: 2025-11-27 09:06:30.815819189 +0000 UTC m=+0.161237723 container exec_died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ceilometer-ipmi-container, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_ipmi, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, version=17.1.12, architecture=x86_64, batch=17.1_20251118.1, build-date=2025-11-19T00:12:45Z, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, managed_by=tripleo_ansible, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:06:30 localhost podman[108379]: unhealthy Nov 27 04:06:30 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:30 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:06:30 localhost podman[108380]: 2025-11-27 09:06:30.843908665 +0000 UTC m=+0.185501196 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 cron, batch=17.1_20251118.1, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, io.buildah.version=1.41.4, container_name=logrotate_crond, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., version=17.1.12, release=1761123044, architecture=x86_64, build-date=2025-11-18T22:49:32Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, name=rhosp17/openstack-cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, distribution-scope=public) Nov 27 04:06:30 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:06:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54727 DF PROTO=TCP SPT=34390 DPT=9105 SEQ=962697044 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AD21A0000000001030307) Nov 27 04:06:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:06:33 localhost systemd[1]: tmp-crun.topCCn.mount: Deactivated successfully. Nov 27 04:06:33 localhost podman[108437]: 2025-11-27 09:06:33.989988258 +0000 UTC m=+0.085554465 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, architecture=x86_64, batch=17.1_20251118.1, config_id=tripleo_step4, release=1761123044, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, distribution-scope=public, io.openshift.expose-services=) Nov 27 04:06:34 localhost podman[108437]: 2025-11-27 09:06:34.351002331 +0000 UTC m=+0.446568538 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, vendor=Red Hat, Inc., batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, distribution-scope=public, name=rhosp17/openstack-nova-compute, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, tcib_managed=true, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, com.redhat.component=openstack-nova-compute-container) Nov 27 04:06:34 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:06:34 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:06:34 localhost recover_tripleo_nova_virtqemud[108461]: 63639 Nov 27 04:06:34 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:06:34 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:06:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:06:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:06:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:06:36 localhost podman[108462]: 2025-11-27 09:06:36.747074454 +0000 UTC m=+0.088346311 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, url=https://www.redhat.com, vendor=Red Hat, Inc., vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, io.openshift.expose-services=, config_id=tripleo_step4) Nov 27 04:06:36 localhost podman[108462]: 2025-11-27 09:06:36.766046565 +0000 UTC m=+0.107318492 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, release=1761123044, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, io.buildah.version=1.41.4, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, build-date=2025-11-19T00:14:25Z, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, architecture=x86_64, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible) Nov 27 04:06:36 localhost podman[108463]: 2025-11-27 09:06:36.803198416 +0000 UTC m=+0.139779646 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, architecture=x86_64, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, release=1761123044, distribution-scope=public, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, build-date=2025-11-18T23:34:05Z, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, vendor=Red Hat, Inc., batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, managed_by=tripleo_ansible) Nov 27 04:06:36 localhost podman[108462]: unhealthy Nov 27 04:06:36 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:36 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:06:36 localhost podman[108464]: 2025-11-27 09:06:36.868782802 +0000 UTC m=+0.202683300 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step5, version=17.1.12, container_name=nova_compute, io.openshift.expose-services=, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, description=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, vcs-type=git, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4) Nov 27 04:06:36 localhost podman[108464]: 2025-11-27 09:06:36.888122943 +0000 UTC m=+0.222023411 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, io.openshift.expose-services=, distribution-scope=public, name=rhosp17/openstack-nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step5, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:36:58Z, com.redhat.component=openstack-nova-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:06:36 localhost podman[108464]: unhealthy Nov 27 04:06:36 localhost podman[108463]: 2025-11-27 09:06:36.896267992 +0000 UTC m=+0.232849192 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public, config_id=tripleo_step4, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.buildah.version=1.41.4, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, managed_by=tripleo_ansible, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc.) Nov 27 04:06:36 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:36 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:06:36 localhost podman[108463]: unhealthy Nov 27 04:06:36 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:06:36 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:06:37 localhost podman[108075]: time="2025-11-27T09:06:37Z" level=warning msg="StopSignal SIGTERM failed to stop container ceilometer_agent_compute in 42 seconds, resorting to SIGKILL" Nov 27 04:06:37 localhost systemd[1]: libpod-f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.scope: Deactivated successfully. Nov 27 04:06:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:d6:23:b6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.110 DST=192.168.122.108 LEN=40 TOS=0x00 PREC=0x00 TTL=64 ID=0 DF PROTO=TCP SPT=6379 DPT=41406 SEQ=3081345371 ACK=0 WINDOW=0 RES=0x00 RST URGP=0 Nov 27 04:06:37 localhost systemd[1]: libpod-f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.scope: Consumed 6.441s CPU time. Nov 27 04:06:37 localhost podman[108075]: 2025-11-27 09:06:37.168311219 +0000 UTC m=+42.094333041 container died f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, batch=17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, managed_by=tripleo_ansible, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ceilometer-compute) Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.timer: Deactivated successfully. Nov 27 04:06:37 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a. Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed to open /run/systemd/transient/f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: No such file or directory Nov 27 04:06:37 localhost podman[108075]: 2025-11-27 09:06:37.230413341 +0000 UTC m=+42.156435163 container cleanup f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-type=git, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, container_name=ceilometer_agent_compute, name=rhosp17/openstack-ceilometer-compute, com.redhat.component=openstack-ceilometer-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, tcib_managed=true, vendor=Red Hat, Inc., config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, release=1761123044, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, managed_by=tripleo_ansible, distribution-scope=public, url=https://www.redhat.com, config_id=tripleo_step4, io.openshift.expose-services=, build-date=2025-11-19T00:11:48Z, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676) Nov 27 04:06:37 localhost podman[108075]: ceilometer_agent_compute Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.timer: Failed to open /run/systemd/transient/f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.timer: No such file or directory Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed to open /run/systemd/transient/f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: No such file or directory Nov 27 04:06:37 localhost podman[108523]: 2025-11-27 09:06:37.268687832 +0000 UTC m=+0.087050565 container cleanup f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=openstack-ceilometer-compute-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-type=git, build-date=2025-11-19T00:11:48Z, tcib_managed=true, architecture=x86_64, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.buildah.version=1.41.4, batch=17.1_20251118.1, url=https://www.redhat.com, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, distribution-scope=public, name=rhosp17/openstack-ceilometer-compute, config_id=tripleo_step4, container_name=ceilometer_agent_compute, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible) Nov 27 04:06:37 localhost systemd[1]: libpod-conmon-f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.scope: Deactivated successfully. Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.timer: Failed to open /run/systemd/transient/f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.timer: No such file or directory Nov 27 04:06:37 localhost systemd[1]: f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: Failed to open /run/systemd/transient/f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a.service: No such file or directory Nov 27 04:06:37 localhost podman[108535]: 2025-11-27 09:06:37.381703016 +0000 UTC m=+0.077779246 container cleanup f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vendor=Red Hat, Inc., io.openshift.expose-services=, tcib_managed=true, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, release=1761123044, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-compute-container, description=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, version=17.1.12, container_name=ceilometer_agent_compute, io.buildah.version=1.41.4, name=rhosp17/openstack-ceilometer-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, batch=17.1_20251118.1, config_id=tripleo_step4) Nov 27 04:06:37 localhost podman[108535]: ceilometer_agent_compute Nov 27 04:06:37 localhost systemd[1]: tripleo_ceilometer_agent_compute.service: Deactivated successfully. Nov 27 04:06:37 localhost systemd[1]: Stopped ceilometer_agent_compute container. Nov 27 04:06:37 localhost systemd[1]: tripleo_ceilometer_agent_compute.service: Consumed 1.100s CPU time, no IO. Nov 27 04:06:37 localhost systemd[1]: var-lib-containers-storage-overlay-3ed13769fb4ebcfcdb2db8fb9f0b5f0e5ef510466865cb9e3869aeb9bf07dd72-merged.mount: Deactivated successfully. Nov 27 04:06:37 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a-userdata-shm.mount: Deactivated successfully. Nov 27 04:06:38 localhost python3.9[108638]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_ceilometer_agent_ipmi.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:06:38 localhost systemd[1]: Reloading. Nov 27 04:06:38 localhost systemd-sysv-generator[108665]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:06:38 localhost systemd-rc-local-generator[108662]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:06:38 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:06:38 localhost systemd[1]: Stopping ceilometer_agent_ipmi container... Nov 27 04:06:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8710 DF PROTO=TCP SPT=50846 DPT=9882 SEQ=111986412 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AE8F70000000001030307) Nov 27 04:06:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8712 DF PROTO=TCP SPT=50846 DPT=9882 SEQ=111986412 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127AF51A0000000001030307) Nov 27 04:06:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:06:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:06:44 localhost systemd[1]: tmp-crun.iKVtK1.mount: Deactivated successfully. Nov 27 04:06:44 localhost podman[108692]: 2025-11-27 09:06:44.248420703 +0000 UTC m=+0.094403663 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, managed_by=tripleo_ansible, batch=17.1_20251118.1, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, summary=Red Hat OpenStack Platform 17.1 collectd, config_id=tripleo_step3, distribution-scope=public) Nov 27 04:06:44 localhost podman[108692]: 2025-11-27 09:06:44.265491972 +0000 UTC m=+0.111474952 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, name=rhosp17/openstack-collectd, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, container_name=collectd, config_id=tripleo_step3, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, tcib_managed=true, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, version=17.1.12, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, url=https://www.redhat.com, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.buildah.version=1.41.4, build-date=2025-11-18T22:51:28Z, io.openshift.expose-services=, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 collectd) Nov 27 04:06:44 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:06:44 localhost podman[108693]: 2025-11-27 09:06:44.338719546 +0000 UTC m=+0.180550575 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-iscsid-container, vcs-type=git, url=https://www.redhat.com, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, distribution-scope=public, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 iscsid, container_name=iscsid, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, batch=17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044) Nov 27 04:06:44 localhost podman[108693]: 2025-11-27 09:06:44.352842986 +0000 UTC m=+0.194674065 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, container_name=iscsid, tcib_managed=true, architecture=x86_64, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, io.openshift.expose-services=, name=rhosp17/openstack-iscsid, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, com.redhat.component=openstack-iscsid-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, version=17.1.12, distribution-scope=public, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, maintainer=OpenStack TripleO Team, vcs-type=git, release=1761123044) Nov 27 04:06:44 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:06:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4891 DF PROTO=TCP SPT=54232 DPT=9102 SEQ=243621658 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B03890000000001030307) Nov 27 04:06:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48308 DF PROTO=TCP SPT=55460 DPT=9100 SEQ=1826708337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B0B9A0000000001030307) Nov 27 04:06:50 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:d6:23:b6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.110 DST=192.168.122.108 LEN=40 TOS=0x00 PREC=0x00 TTL=64 ID=0 DF PROTO=TCP SPT=6379 DPT=41406 SEQ=3081345371 ACK=0 WINDOW=0 RES=0x00 RST URGP=0 Nov 27 04:06:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=26959 DF PROTO=TCP SPT=32802 DPT=9101 SEQ=2044993144 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B235A0000000001030307) Nov 27 04:06:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:06:54 localhost podman[108729]: 2025-11-27 09:06:54.98693947 +0000 UTC m=+0.081321691 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, version=17.1.12, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vendor=Red Hat, Inc., vcs-type=git, com.redhat.component=openstack-qdrouterd-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:46Z, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, managed_by=tripleo_ansible, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, container_name=metrics_qdr, config_id=tripleo_step1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, description=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd) Nov 27 04:06:55 localhost podman[108729]: 2025-11-27 09:06:55.21012103 +0000 UTC m=+0.304503291 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.buildah.version=1.41.4, batch=17.1_20251118.1, build-date=2025-11-18T22:49:46Z, version=17.1.12, container_name=metrics_qdr, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, io.openshift.expose-services=, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, release=1761123044, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, name=rhosp17/openstack-qdrouterd) Nov 27 04:06:55 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:06:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36307 DF PROTO=TCP SPT=55842 DPT=9105 SEQ=3820376470 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B2F9A0000000001030307) Nov 27 04:06:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=26961 DF PROTO=TCP SPT=32802 DPT=9101 SEQ=2044993144 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B3B1B0000000001030307) Nov 27 04:07:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:07:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:07:00 localhost podman[108759]: Error: container ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 is not running Nov 27 04:07:00 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Main process exited, code=exited, status=125/n/a Nov 27 04:07:00 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed with result 'exit-code'. Nov 27 04:07:01 localhost systemd[1]: tmp-crun.VuW5sU.mount: Deactivated successfully. Nov 27 04:07:01 localhost podman[108760]: 2025-11-27 09:07:01.039215762 +0000 UTC m=+0.131831102 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, release=1761123044, name=rhosp17/openstack-cron, url=https://www.redhat.com, batch=17.1_20251118.1, container_name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step4, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, build-date=2025-11-18T22:49:32Z, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 cron, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.buildah.version=1.41.4) Nov 27 04:07:01 localhost podman[108760]: 2025-11-27 09:07:01.053856416 +0000 UTC m=+0.146471766 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:49:32Z, release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, tcib_managed=true, container_name=logrotate_crond, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, architecture=x86_64, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, vendor=Red Hat, Inc., com.redhat.component=openstack-cron-container, version=17.1.12, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 04:07:01 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:07:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36309 DF PROTO=TCP SPT=55842 DPT=9105 SEQ=3820376470 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B475A0000000001030307) Nov 27 04:07:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:07:04 localhost podman[108790]: 2025-11-27 09:07:04.744271198 +0000 UTC m=+0.085385971 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, config_id=tripleo_step4, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, vendor=Red Hat, Inc., version=17.1.12, url=https://www.redhat.com, io.buildah.version=1.41.4, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, batch=17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible) Nov 27 04:07:05 localhost podman[108790]: 2025-11-27 09:07:05.123977114 +0000 UTC m=+0.465091877 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, batch=17.1_20251118.1, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, version=17.1.12, name=rhosp17/openstack-nova-compute, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_migration_target, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, release=1761123044, config_id=tripleo_step4, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., tcib_managed=true, com.redhat.component=openstack-nova-compute-container, summary=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:07:05 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:07:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:07:06 localhost systemd[1]: tmp-crun.nFm80c.mount: Deactivated successfully. Nov 27 04:07:07 localhost podman[108814]: 2025-11-27 09:07:06.99626376 +0000 UTC m=+0.087929510 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.buildah.version=1.41.4, vendor=Red Hat, Inc., distribution-scope=public, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 04:07:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:07:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:07:07 localhost podman[108814]: 2025-11-27 09:07:07.041991932 +0000 UTC m=+0.133657702 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, io.openshift.expose-services=, container_name=ovn_metadata_agent, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:07:07 localhost podman[108814]: unhealthy Nov 27 04:07:07 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:07 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:07:07 localhost systemd[1]: tmp-crun.12DTgz.mount: Deactivated successfully. Nov 27 04:07:07 localhost podman[108833]: 2025-11-27 09:07:07.099193172 +0000 UTC m=+0.073821079 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, config_id=tripleo_step5, vcs-type=git, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com) Nov 27 04:07:07 localhost podman[108833]: 2025-11-27 09:07:07.146917847 +0000 UTC m=+0.121545824 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, distribution-scope=public, vcs-type=git, url=https://www.redhat.com, name=rhosp17/openstack-nova-compute, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, managed_by=tripleo_ansible, version=17.1.12, config_id=tripleo_step5, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute) Nov 27 04:07:07 localhost podman[108833]: unhealthy Nov 27 04:07:07 localhost podman[108832]: 2025-11-27 09:07:07.15703081 +0000 UTC m=+0.133858886 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, config_id=tripleo_step4, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, io.buildah.version=1.41.4, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, architecture=x86_64, release=1761123044, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 04:07:07 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:07 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:07:07 localhost podman[108832]: 2025-11-27 09:07:07.200220313 +0000 UTC m=+0.177048349 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, vcs-type=git, url=https://www.redhat.com, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, name=rhosp17/openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, vendor=Red Hat, Inc., tcib_managed=true, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, container_name=ovn_controller, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container) Nov 27 04:07:07 localhost podman[108832]: unhealthy Nov 27 04:07:07 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:07 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:07:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=26962 DF PROTO=TCP SPT=32802 DPT=9101 SEQ=2044993144 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B5BDA0000000001030307) Nov 27 04:07:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2756 DF PROTO=TCP SPT=34278 DPT=9882 SEQ=1844218617 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B5E260000000001030307) Nov 27 04:07:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2758 DF PROTO=TCP SPT=34278 DPT=9882 SEQ=1844218617 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B6A1B0000000001030307) Nov 27 04:07:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:07:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:07:14 localhost podman[108873]: 2025-11-27 09:07:14.747762458 +0000 UTC m=+0.092984606 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, build-date=2025-11-18T23:44:13Z, url=https://www.redhat.com, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, version=17.1.12, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, batch=17.1_20251118.1, architecture=x86_64, distribution-scope=public, name=rhosp17/openstack-iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, vendor=Red Hat, Inc.) Nov 27 04:07:14 localhost podman[108872]: 2025-11-27 09:07:14.785234327 +0000 UTC m=+0.132948781 container health_status 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, health_status=healthy, io.buildah.version=1.41.4, com.redhat.component=openstack-collectd-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, container_name=collectd, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 collectd, release=1761123044, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:07:14 localhost podman[108872]: 2025-11-27 09:07:14.799815649 +0000 UTC m=+0.147530083 container exec_died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, config_id=tripleo_step3, architecture=x86_64, io.openshift.expose-services=, io.buildah.version=1.41.4, name=rhosp17/openstack-collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, url=https://www.redhat.com, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T22:51:28Z, com.redhat.component=openstack-collectd-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, managed_by=tripleo_ansible, vendor=Red Hat, Inc., org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, batch=17.1_20251118.1, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, maintainer=OpenStack TripleO Team, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a) Nov 27 04:07:14 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Deactivated successfully. Nov 27 04:07:14 localhost podman[108873]: 2025-11-27 09:07:14.85773897 +0000 UTC m=+0.202961128 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.buildah.version=1.41.4, architecture=x86_64, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, managed_by=tripleo_ansible, release=1761123044, summary=Red Hat OpenStack Platform 17.1 iscsid, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, build-date=2025-11-18T23:44:13Z, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, distribution-scope=public, name=rhosp17/openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-iscsid-container, container_name=iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.expose-services=, version=17.1.12) Nov 27 04:07:14 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:07:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22182 DF PROTO=TCP SPT=42404 DPT=9102 SEQ=2690712891 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B78BA0000000001030307) Nov 27 04:07:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43656 DF PROTO=TCP SPT=44598 DPT=9100 SEQ=4151292011 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B80DA0000000001030307) Nov 27 04:07:20 localhost podman[108679]: time="2025-11-27T09:07:20Z" level=warning msg="StopSignal SIGTERM failed to stop container ceilometer_agent_ipmi in 42 seconds, resorting to SIGKILL" Nov 27 04:07:20 localhost systemd[1]: libpod-ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.scope: Deactivated successfully. Nov 27 04:07:20 localhost systemd[1]: libpod-ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.scope: Consumed 6.580s CPU time. Nov 27 04:07:20 localhost podman[108679]: 2025-11-27 09:07:20.703180282 +0000 UTC m=+42.104924982 container died ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, name=rhosp17/openstack-ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, architecture=x86_64, version=17.1.12, vendor=Red Hat, Inc., managed_by=tripleo_ansible, tcib_managed=true, distribution-scope=public, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, build-date=2025-11-19T00:12:45Z, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=) Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.timer: Deactivated successfully. Nov 27 04:07:20 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044. Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed to open /run/systemd/transient/ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: No such file or directory Nov 27 04:07:20 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044-userdata-shm.mount: Deactivated successfully. Nov 27 04:07:20 localhost podman[108679]: 2025-11-27 09:07:20.776331223 +0000 UTC m=+42.178075923 container cleanup ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, name=rhosp17/openstack-ceilometer-ipmi, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, io.buildah.version=1.41.4, tcib_managed=true, version=17.1.12, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, url=https://www.redhat.com, config_id=tripleo_step4, container_name=ceilometer_agent_ipmi, build-date=2025-11-19T00:12:45Z, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, vcs-type=git, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:07:20 localhost podman[108679]: ceilometer_agent_ipmi Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.timer: Failed to open /run/systemd/transient/ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.timer: No such file or directory Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed to open /run/systemd/transient/ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: No such file or directory Nov 27 04:07:20 localhost podman[108974]: 2025-11-27 09:07:20.850633524 +0000 UTC m=+0.136301923 container cleanup ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, com.redhat.component=openstack-ceilometer-ipmi-container, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, build-date=2025-11-19T00:12:45Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, url=https://www.redhat.com, distribution-scope=public, name=rhosp17/openstack-ceilometer-ipmi, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, architecture=x86_64, vendor=Red Hat, Inc., tcib_managed=true, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, release=1761123044, container_name=ceilometer_agent_ipmi, maintainer=OpenStack TripleO Team, config_id=tripleo_step4) Nov 27 04:07:20 localhost systemd[1]: libpod-conmon-ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.scope: Deactivated successfully. Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.timer: Failed to open /run/systemd/transient/ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.timer: No such file or directory Nov 27 04:07:20 localhost systemd[1]: ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: Failed to open /run/systemd/transient/ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044.service: No such file or directory Nov 27 04:07:20 localhost podman[108991]: 2025-11-27 09:07:20.960859212 +0000 UTC m=+0.077645492 container cleanup ca93886e8cc37776f0017b597e04a0d38077b54f5322a77227b2a62556c87044 (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1, name=ceilometer_agent_ipmi, name=rhosp17/openstack-ceilometer-ipmi, version=17.1.12, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-ipmi, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:12:45Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-ipmi:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer-agent-ipmi.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, config_id=tripleo_step4, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ceilometer-ipmi-container, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-ipmi, distribution-scope=public, release=1761123044, managed_by=tripleo_ansible, io.openshift.expose-services=, io.buildah.version=1.41.4, container_name=ceilometer_agent_ipmi) Nov 27 04:07:20 localhost podman[108991]: ceilometer_agent_ipmi Nov 27 04:07:20 localhost systemd[1]: tripleo_ceilometer_agent_ipmi.service: Deactivated successfully. Nov 27 04:07:20 localhost systemd[1]: Stopped ceilometer_agent_ipmi container. Nov 27 04:07:21 localhost python3.9[109110]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_collectd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43657 DF PROTO=TCP SPT=44598 DPT=9100 SEQ=4151292011 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B909A0000000001030307) Nov 27 04:07:21 localhost systemd[1]: var-lib-containers-storage-overlay-e7f27f4aecc7722c6029129a1910e39be572d823c20ebf6e60c902dacc886f74-merged.mount: Deactivated successfully. Nov 27 04:07:21 localhost systemd[1]: Reloading. Nov 27 04:07:21 localhost systemd-rc-local-generator[109138]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:07:21 localhost systemd-sysv-generator[109143]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:07:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:07:22 localhost systemd[1]: Stopping collectd container... Nov 27 04:07:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29117 DF PROTO=TCP SPT=58994 DPT=9101 SEQ=659313165 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127B989A0000000001030307) Nov 27 04:07:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:07:25 localhost systemd[1]: tmp-crun.dmupNF.mount: Deactivated successfully. Nov 27 04:07:25 localhost podman[109165]: 2025-11-27 09:07:25.744433167 +0000 UTC m=+0.091933757 container health_status 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, health_status=healthy, container_name=metrics_qdr, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64, version=17.1.12, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, vcs-type=git, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, distribution-scope=public, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, managed_by=tripleo_ansible, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, com.redhat.component=openstack-qdrouterd-container, config_id=tripleo_step1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}) Nov 27 04:07:25 localhost podman[109165]: 2025-11-27 09:07:25.937052885 +0000 UTC m=+0.284553555 container exec_died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, build-date=2025-11-18T22:49:46Z, managed_by=tripleo_ansible, release=1761123044, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, vendor=Red Hat, Inc., container_name=metrics_qdr, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step1, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, io.buildah.version=1.41.4, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, url=https://www.redhat.com, name=rhosp17/openstack-qdrouterd) Nov 27 04:07:25 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Deactivated successfully. Nov 27 04:07:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47886 DF PROTO=TCP SPT=55916 DPT=9101 SEQ=2369423777 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BA3DA0000000001030307) Nov 27 04:07:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54730 DF PROTO=TCP SPT=34390 DPT=9105 SEQ=962697044 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BAFDA0000000001030307) Nov 27 04:07:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:07:31 localhost systemd[1]: tmp-crun.HFAuCi.mount: Deactivated successfully. Nov 27 04:07:31 localhost podman[109194]: 2025-11-27 09:07:31.499848706 +0000 UTC m=+0.094901617 container health_status e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, health_status=healthy, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, url=https://www.redhat.com, managed_by=tripleo_ansible, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, name=rhosp17/openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.component=openstack-cron-container, vendor=Red Hat, Inc., version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, batch=17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:32Z, container_name=logrotate_crond, release=1761123044) Nov 27 04:07:31 localhost podman[109194]: 2025-11-27 09:07:31.51188579 +0000 UTC m=+0.106938681 container exec_died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-cron-container, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, release=1761123044, url=https://www.redhat.com, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=logrotate_crond, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, managed_by=tripleo_ansible, vendor=Red Hat, Inc., name=rhosp17/openstack-cron, build-date=2025-11-18T22:49:32Z, summary=Red Hat OpenStack Platform 17.1 cron, vcs-type=git, io.openshift.expose-services=, batch=17.1_20251118.1, io.buildah.version=1.41.4) Nov 27 04:07:31 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Deactivated successfully. Nov 27 04:07:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=61506 DF PROTO=TCP SPT=54150 DPT=9105 SEQ=1683592853 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BBC5A0000000001030307) Nov 27 04:07:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:07:35 localhost systemd[1]: tmp-crun.ZyHvaK.mount: Deactivated successfully. Nov 27 04:07:35 localhost podman[109214]: 2025-11-27 09:07:35.749775148 +0000 UTC m=+0.093299655 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, version=17.1.12, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_migration_target, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, vcs-type=git, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, io.openshift.expose-services=, managed_by=tripleo_ansible, release=1761123044, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:07:36 localhost podman[109214]: 2025-11-27 09:07:36.131094868 +0000 UTC m=+0.474619395 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, distribution-scope=public, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, vendor=Red Hat, Inc., vcs-type=git, container_name=nova_migration_target, managed_by=tripleo_ansible, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-compute-container, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.expose-services=, architecture=x86_64) Nov 27 04:07:36 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:07:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29120 DF PROTO=TCP SPT=58994 DPT=9101 SEQ=659313165 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BCFDB0000000001030307) Nov 27 04:07:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:07:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:07:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:07:37 localhost systemd[1]: tmp-crun.W1GJlP.mount: Deactivated successfully. Nov 27 04:07:38 localhost podman[109238]: 2025-11-27 09:07:38.003344032 +0000 UTC m=+0.091882226 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, container_name=ovn_controller, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-ovn-controller-container, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, tcib_managed=true, url=https://www.redhat.com, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}) Nov 27 04:07:38 localhost podman[109238]: 2025-11-27 09:07:38.023867934 +0000 UTC m=+0.112406148 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-18T23:34:05Z, architecture=x86_64, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, container_name=ovn_controller, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:07:38 localhost podman[109238]: unhealthy Nov 27 04:07:38 localhost systemd[1]: tmp-crun.SDlea8.mount: Deactivated successfully. Nov 27 04:07:38 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:38 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:07:38 localhost podman[109237]: 2025-11-27 09:07:38.045563669 +0000 UTC m=+0.136195759 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-type=git, tcib_managed=true, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, architecture=x86_64, io.openshift.expose-services=, version=17.1.12, batch=17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044) Nov 27 04:07:38 localhost podman[109237]: 2025-11-27 09:07:38.090009866 +0000 UTC m=+0.180641946 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, tcib_managed=true, distribution-scope=public, io.buildah.version=1.41.4, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, config_id=tripleo_step4, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, url=https://www.redhat.com, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}) Nov 27 04:07:38 localhost podman[109237]: unhealthy Nov 27 04:07:38 localhost podman[109239]: 2025-11-27 09:07:38.100597051 +0000 UTC m=+0.185750464 container health_status ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, health_status=unhealthy, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, build-date=2025-11-19T00:36:58Z, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, description=Red Hat OpenStack Platform 17.1 nova-compute, vendor=Red Hat, Inc., vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, url=https://www.redhat.com, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, container_name=nova_compute, tcib_managed=true, name=rhosp17/openstack-nova-compute, config_id=tripleo_step5, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4) Nov 27 04:07:38 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:38 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:07:38 localhost podman[109239]: 2025-11-27 09:07:38.122909142 +0000 UTC m=+0.208062565 container exec_died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., distribution-scope=public, architecture=x86_64, container_name=nova_compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-compute, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, com.redhat.component=openstack-nova-compute-container, release=1761123044, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, build-date=2025-11-19T00:36:58Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, config_id=tripleo_step5, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, batch=17.1_20251118.1, name=rhosp17/openstack-nova-compute) Nov 27 04:07:38 localhost podman[109239]: unhealthy Nov 27 04:07:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:07:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40089 DF PROTO=TCP SPT=56388 DPT=9882 SEQ=3914016577 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BD3570000000001030307) Nov 27 04:07:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40091 DF PROTO=TCP SPT=56388 DPT=9882 SEQ=3914016577 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BDF5A0000000001030307) Nov 27 04:07:44 localhost systemd[1]: libpod-193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.scope: Deactivated successfully. Nov 27 04:07:44 localhost systemd[1]: libpod-193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.scope: Consumed 2.229s CPU time. Nov 27 04:07:44 localhost podman[109152]: 2025-11-27 09:07:44.788987686 +0000 UTC m=+22.723446182 container died 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, com.redhat.component=openstack-collectd-container, managed_by=tripleo_ansible, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, distribution-scope=public, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:51:28Z, url=https://www.redhat.com, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, version=17.1.12, release=1761123044, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, vcs-type=git, name=rhosp17/openstack-collectd, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, description=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true) Nov 27 04:07:44 localhost systemd[1]: tmp-crun.BBKno9.mount: Deactivated successfully. Nov 27 04:07:44 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.timer: Deactivated successfully. Nov 27 04:07:44 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23. Nov 27 04:07:44 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Failed to open /run/systemd/transient/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: No such file or directory Nov 27 04:07:44 localhost podman[109152]: 2025-11-27 09:07:44.840682029 +0000 UTC m=+22.775140515 container cleanup 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, version=17.1.12, summary=Red Hat OpenStack Platform 17.1 collectd, io.buildah.version=1.41.4, batch=17.1_20251118.1, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, release=1761123044, com.redhat.component=openstack-collectd-container, build-date=2025-11-18T22:51:28Z, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, name=rhosp17/openstack-collectd, container_name=collectd, architecture=x86_64, vcs-type=git, tcib_managed=true, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, config_id=tripleo_step3, vendor=Red Hat, Inc., io.openshift.expose-services=, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, description=Red Hat OpenStack Platform 17.1 collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:07:44 localhost podman[109152]: collectd Nov 27 04:07:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:07:44 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.timer: Failed to open /run/systemd/transient/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.timer: No such file or directory Nov 27 04:07:44 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Failed to open /run/systemd/transient/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: No such file or directory Nov 27 04:07:44 localhost podman[109299]: 2025-11-27 09:07:44.889169545 +0000 UTC m=+0.096237133 container cleanup 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, summary=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, managed_by=tripleo_ansible, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, container_name=collectd, distribution-scope=public, architecture=x86_64, name=rhosp17/openstack-collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, io.openshift.expose-services=, build-date=2025-11-18T22:51:28Z, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 collectd, com.redhat.component=openstack-collectd-container, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044) Nov 27 04:07:44 localhost systemd[1]: tripleo_collectd.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:07:44 localhost systemd[1]: libpod-conmon-193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.scope: Deactivated successfully. Nov 27 04:07:44 localhost podman[109315]: 2025-11-27 09:07:44.984688187 +0000 UTC m=+0.087258941 container health_status 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, health_status=healthy, container_name=iscsid, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vendor=Red Hat, Inc., build-date=2025-11-18T23:44:13Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, maintainer=OpenStack TripleO Team, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, vcs-type=git, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, com.redhat.component=openstack-iscsid-container, release=1761123044, config_id=tripleo_step3, io.openshift.expose-services=, managed_by=tripleo_ansible, architecture=x86_64, url=https://www.redhat.com, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 04:07:44 localhost podman[109315]: 2025-11-27 09:07:44.99814213 +0000 UTC m=+0.100712934 container exec_died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, config_id=tripleo_step3, container_name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, com.redhat.component=openstack-iscsid-container, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, io.buildah.version=1.41.4, batch=17.1_20251118.1, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, name=rhosp17/openstack-iscsid) Nov 27 04:07:45 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Deactivated successfully. Nov 27 04:07:45 localhost podman[109347]: error opening file `/run/crun/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23/status`: No such file or directory Nov 27 04:07:45 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.timer: Failed to open /run/systemd/transient/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.timer: No such file or directory Nov 27 04:07:45 localhost systemd[1]: 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: Failed to open /run/systemd/transient/193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23.service: No such file or directory Nov 27 04:07:45 localhost podman[109320]: 2025-11-27 09:07:45.086107298 +0000 UTC m=+0.178050115 container cleanup 193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23 (image=registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1, name=collectd, version=17.1.12, description=Red Hat OpenStack Platform 17.1 collectd, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 collectd, config_data={'cap_add': ['IPC_LOCK'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd31718fcd17fdeee6489534105191c7a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-collectd:17.1', 'memory': '512m', 'net': 'host', 'pid': 'host', 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/collectd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/containers/storage/overlay-containers:/var/lib/containers/storage/overlay-containers:ro', '/var/lib/config-data/puppet-generated/collectd:/var/lib/kolla/config_files/src:ro', '/var/log/containers/collectd:/var/log/collectd:rw,z', '/var/lib/container-config-scripts:/config-scripts:ro', '/var/lib/container-user-scripts:/scripts:z', '/run:/run:rw', '/sys/fs/cgroup:/sys/fs/cgroup:ro']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=collectd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-collectd-container, config_id=tripleo_step3, io.k8s.display-name=Red Hat OpenStack Platform 17.1 collectd, vcs-type=git, build-date=2025-11-18T22:51:28Z, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, name=rhosp17/openstack-collectd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, managed_by=tripleo_ansible, vendor=Red Hat, Inc., tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-collectd, batch=17.1_20251118.1, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 collectd, url=https://www.redhat.com, maintainer=OpenStack TripleO Team) Nov 27 04:07:45 localhost podman[109320]: collectd Nov 27 04:07:45 localhost systemd[1]: tripleo_collectd.service: Failed with result 'exit-code'. Nov 27 04:07:45 localhost systemd[1]: Stopped collectd container. Nov 27 04:07:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60524 DF PROTO=TCP SPT=34024 DPT=9102 SEQ=4260730175 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BEDE80000000001030307) Nov 27 04:07:45 localhost systemd[1]: var-lib-containers-storage-overlay-706e7dad99f72ad0a2bbd0167294cca4c86e34b8636b5f5f9b7b80818f253353-merged.mount: Deactivated successfully. Nov 27 04:07:45 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-193d6a4e7d63419a31b5fd1be86882f6589633af0fbffa9eb00f90867eac5b23-userdata-shm.mount: Deactivated successfully. Nov 27 04:07:45 localhost python3.9[109441]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_iscsid.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:45 localhost systemd[1]: Reloading. Nov 27 04:07:46 localhost systemd-sysv-generator[109472]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:07:46 localhost systemd-rc-local-generator[109467]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:07:46 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:07:46 localhost systemd[1]: Stopping iscsid container... Nov 27 04:07:46 localhost systemd[1]: libpod-84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.scope: Deactivated successfully. Nov 27 04:07:46 localhost systemd[1]: libpod-84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.scope: Consumed 1.180s CPU time. Nov 27 04:07:46 localhost podman[109482]: 2025-11-27 09:07:46.338132599 +0000 UTC m=+0.073881041 container died 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, build-date=2025-11-18T23:44:13Z, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-iscsid, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, com.redhat.component=openstack-iscsid-container, io.openshift.expose-services=, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-type=git, io.buildah.version=1.41.4, release=1761123044, architecture=x86_64, container_name=iscsid, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, config_id=tripleo_step3, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, version=17.1.12, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d) Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.timer: Deactivated successfully. Nov 27 04:07:46 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9. Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Failed to open /run/systemd/transient/84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: No such file or directory Nov 27 04:07:46 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9-userdata-shm.mount: Deactivated successfully. Nov 27 04:07:46 localhost podman[109482]: 2025-11-27 09:07:46.384148589 +0000 UTC m=+0.119897031 container cleanup 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, tcib_managed=true, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, vcs-type=git, name=rhosp17/openstack-iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, io.buildah.version=1.41.4, config_id=tripleo_step3, build-date=2025-11-18T23:44:13Z, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, com.redhat.component=openstack-iscsid-container, vendor=Red Hat, Inc., release=1761123044, version=17.1.12, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, container_name=iscsid, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:07:46 localhost podman[109482]: iscsid Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.timer: Failed to open /run/systemd/transient/84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.timer: No such file or directory Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Failed to open /run/systemd/transient/84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: No such file or directory Nov 27 04:07:46 localhost podman[109495]: 2025-11-27 09:07:46.415336699 +0000 UTC m=+0.066442881 container cleanup 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-iscsid-container, vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-iscsid, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, managed_by=tripleo_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, summary=Red Hat OpenStack Platform 17.1 iscsid, url=https://www.redhat.com, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, release=1761123044, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:44:13Z, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, distribution-scope=public, container_name=iscsid, io.openshift.expose-services=, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3) Nov 27 04:07:46 localhost systemd[1]: libpod-conmon-84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.scope: Deactivated successfully. Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.timer: Failed to open /run/systemd/transient/84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.timer: No such file or directory Nov 27 04:07:46 localhost systemd[1]: 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: Failed to open /run/systemd/transient/84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9.service: No such file or directory Nov 27 04:07:46 localhost podman[109511]: 2025-11-27 09:07:46.511003755 +0000 UTC m=+0.066844361 container cleanup 84514c76d66ee0d36dd7d32c25663ceb43b4fe0e3271ea4320b98df99cd866f9 (image=registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1, name=iscsid, io.buildah.version=1.41.4, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 iscsid, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, build-date=2025-11-18T23:44:13Z, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, managed_by=tripleo_ansible, config_id=tripleo_step3, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 iscsid, description=Red Hat OpenStack Platform 17.1 iscsid, summary=Red Hat OpenStack Platform 17.1 iscsid, tcib_managed=true, com.redhat.component=openstack-iscsid-container, release=1761123044, maintainer=OpenStack TripleO Team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-iscsid:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 2, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/iscsid.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run:/run', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/etc/target:/etc/target:z', '/var/lib/iscsi:/var/lib/iscsi:z']}, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-iscsid, io.openshift.expose-services=, org.opencontainers.image.revision=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, url=https://www.redhat.com, container_name=iscsid, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=5714445d3136fb8f8cd5e0726e4e3e709c68ad0d, name=rhosp17/openstack-iscsid) Nov 27 04:07:46 localhost podman[109511]: iscsid Nov 27 04:07:46 localhost systemd[1]: tripleo_iscsid.service: Deactivated successfully. Nov 27 04:07:46 localhost systemd[1]: Stopped iscsid container. Nov 27 04:07:46 localhost systemd[1]: var-lib-containers-storage-overlay-47567cd0527d828c782715e735bfac6bc84c5467ad9fa21661ac755910450a2a-merged.mount: Deactivated successfully. Nov 27 04:07:47 localhost python3.9[109613]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_logrotate_crond.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:47 localhost systemd[1]: Reloading. Nov 27 04:07:47 localhost systemd-rc-local-generator[109638]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:07:47 localhost systemd-sysv-generator[109643]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:07:47 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:07:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22187 DF PROTO=TCP SPT=42404 DPT=9102 SEQ=2690712891 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127BF5DA0000000001030307) Nov 27 04:07:47 localhost systemd[1]: Stopping logrotate_crond container... Nov 27 04:07:47 localhost systemd[1]: libpod-e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.scope: Deactivated successfully. Nov 27 04:07:47 localhost systemd[1]: libpod-e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.scope: Consumed 1.109s CPU time. Nov 27 04:07:47 localhost podman[109653]: 2025-11-27 09:07:47.726484031 +0000 UTC m=+0.080114799 container died e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, url=https://www.redhat.com, distribution-scope=public, config_id=tripleo_step4, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, version=17.1.12, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, com.redhat.component=openstack-cron-container, summary=Red Hat OpenStack Platform 17.1 cron, build-date=2025-11-18T22:49:32Z, vcs-type=git, name=rhosp17/openstack-cron, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=logrotate_crond, vendor=Red Hat, Inc., release=1761123044, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.timer: Deactivated successfully. Nov 27 04:07:47 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d. Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Failed to open /run/systemd/transient/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: No such file or directory Nov 27 04:07:47 localhost podman[109653]: 2025-11-27 09:07:47.783905767 +0000 UTC m=+0.137536485 container cleanup e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 cron, distribution-scope=public, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, managed_by=tripleo_ansible, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1, release=1761123044, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, build-date=2025-11-18T22:49:32Z, vendor=Red Hat, Inc., name=rhosp17/openstack-cron, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, container_name=logrotate_crond, url=https://www.redhat.com, com.redhat.component=openstack-cron-container, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, description=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:07:47 localhost podman[109653]: logrotate_crond Nov 27 04:07:47 localhost systemd[1]: var-lib-containers-storage-overlay-6fdd4104eccc8c43c3ebaef353a49eae132b42626593a557e6485db08bad4f8e-merged.mount: Deactivated successfully. Nov 27 04:07:47 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d-userdata-shm.mount: Deactivated successfully. Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.timer: Failed to open /run/systemd/transient/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.timer: No such file or directory Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Failed to open /run/systemd/transient/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: No such file or directory Nov 27 04:07:47 localhost podman[109666]: 2025-11-27 09:07:47.827314297 +0000 UTC m=+0.089237085 container cleanup e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, container_name=logrotate_crond, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, description=Red Hat OpenStack Platform 17.1 cron, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, summary=Red Hat OpenStack Platform 17.1 cron, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, version=17.1.12, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-cron-container, io.openshift.expose-services=, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, config_id=tripleo_step4, batch=17.1_20251118.1, io.buildah.version=1.41.4, url=https://www.redhat.com, build-date=2025-11-18T22:49:32Z, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, release=1761123044, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-cron, managed_by=tripleo_ansible, vendor=Red Hat, Inc., vcs-type=git) Nov 27 04:07:47 localhost systemd[1]: libpod-conmon-e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.scope: Deactivated successfully. Nov 27 04:07:47 localhost podman[109695]: error opening file `/run/crun/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d/status`: No such file or directory Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.timer: Failed to open /run/systemd/transient/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.timer: No such file or directory Nov 27 04:07:47 localhost systemd[1]: e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: Failed to open /run/systemd/transient/e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d.service: No such file or directory Nov 27 04:07:47 localhost podman[109684]: 2025-11-27 09:07:47.935960413 +0000 UTC m=+0.078992799 container cleanup e123593b837617cc67cbc1bb038136a2018032e186cd73fb805792c810f7622d (image=registry.redhat.io/rhosp-rhel9/openstack-cron:17.1, name=logrotate_crond, description=Red Hat OpenStack Platform 17.1 cron, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, build-date=2025-11-18T22:49:32Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-cron, name=rhosp17/openstack-cron, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, architecture=x86_64, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vcs-type=git, com.redhat.component=openstack-cron-container, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 cron, version=17.1.12, io.k8s.display-name=Red Hat OpenStack Platform 17.1 cron, container_name=logrotate_crond, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '53ed83bb0cae779ff95edb2002262c6f'}, 'healthcheck': {'test': '/usr/share/openstack-tripleo-common/healthcheck/cron'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-cron:17.1', 'net': 'none', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/logrotate-crond.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/crond:/var/lib/kolla/config_files/src:ro', '/var/log/containers:/var/log/containers:z']}, io.openshift.expose-services=, url=https://www.redhat.com, batch=17.1_20251118.1, managed_by=tripleo_ansible, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, release=1761123044, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 cron) Nov 27 04:07:47 localhost podman[109684]: logrotate_crond Nov 27 04:07:47 localhost systemd[1]: tripleo_logrotate_crond.service: Deactivated successfully. Nov 27 04:07:47 localhost systemd[1]: Stopped logrotate_crond container. Nov 27 04:07:48 localhost python3.9[109789]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_metrics_qdr.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:48 localhost systemd[1]: Reloading. Nov 27 04:07:48 localhost systemd-sysv-generator[109817]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:07:48 localhost systemd-rc-local-generator[109812]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:07:48 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:07:49 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:07:49 localhost systemd[1]: Stopping metrics_qdr container... Nov 27 04:07:49 localhost recover_tripleo_nova_virtqemud[109831]: 63639 Nov 27 04:07:49 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:07:49 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:07:49 localhost systemd[1]: tmp-crun.ih7bcI.mount: Deactivated successfully. Nov 27 04:07:49 localhost kernel: qdrouterd[56290]: segfault at 0 ip 00007fd38cc627cb sp 00007ffff0bcc9b0 error 4 in libc.so.6[7fd38cbff000+175000] Nov 27 04:07:49 localhost kernel: Code: 0b 00 64 44 89 23 85 c0 75 d4 e9 2b ff ff ff e8 db a5 00 00 e9 fd fe ff ff e8 41 1d 0d 00 90 f3 0f 1e fa 41 54 55 48 89 fd 53 <8b> 07 f6 c4 20 0f 85 aa 00 00 00 89 c2 81 e2 00 80 00 00 0f 84 a9 Nov 27 04:07:49 localhost systemd[1]: Created slice Slice /system/systemd-coredump. Nov 27 04:07:49 localhost systemd[1]: Started Process Core Dump (PID 109844/UID 0). Nov 27 04:07:49 localhost systemd-coredump[109845]: Resource limits disable core dumping for process 56290 (qdrouterd). Nov 27 04:07:49 localhost systemd-coredump[109845]: Process 56290 (qdrouterd) of user 42465 dumped core. Nov 27 04:07:49 localhost systemd[1]: systemd-coredump@0-109844-0.service: Deactivated successfully. Nov 27 04:07:49 localhost systemd[1]: libpod-646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.scope: Deactivated successfully. Nov 27 04:07:49 localhost podman[109830]: 2025-11-27 09:07:49.294886783 +0000 UTC m=+0.244474837 container died 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, summary=Red Hat OpenStack Platform 17.1 qdrouterd, container_name=metrics_qdr, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-18T22:49:46Z, config_id=tripleo_step1, description=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, managed_by=tripleo_ansible, distribution-scope=public, name=rhosp17/openstack-qdrouterd, version=17.1.12, maintainer=OpenStack TripleO Team, release=1761123044, vcs-type=git, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:07:49 localhost systemd[1]: libpod-646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.scope: Consumed 28.675s CPU time. Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.timer: Deactivated successfully. Nov 27 04:07:49 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00. Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Failed to open /run/systemd/transient/646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: No such file or directory Nov 27 04:07:49 localhost podman[109830]: 2025-11-27 09:07:49.347379176 +0000 UTC m=+0.296967200 container cleanup 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, vendor=Red Hat, Inc., container_name=metrics_qdr, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, version=17.1.12, config_id=tripleo_step1, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, com.redhat.component=openstack-qdrouterd-container, architecture=x86_64, release=1761123044, managed_by=tripleo_ansible, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, maintainer=OpenStack TripleO Team, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, name=rhosp17/openstack-qdrouterd, build-date=2025-11-18T22:49:46Z, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, vcs-type=git) Nov 27 04:07:49 localhost podman[109830]: metrics_qdr Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.timer: Failed to open /run/systemd/transient/646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.timer: No such file or directory Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Failed to open /run/systemd/transient/646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: No such file or directory Nov 27 04:07:49 localhost podman[109849]: 2025-11-27 09:07:49.383489018 +0000 UTC m=+0.081639029 container cleanup 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, tcib_managed=true, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, architecture=x86_64, container_name=metrics_qdr, url=https://www.redhat.com, config_id=tripleo_step1, build-date=2025-11-18T22:49:46Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, batch=17.1_20251118.1, vendor=Red Hat, Inc., config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, managed_by=tripleo_ansible, name=rhosp17/openstack-qdrouterd, summary=Red Hat OpenStack Platform 17.1 qdrouterd, vcs-type=git, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, version=17.1.12, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd) Nov 27 04:07:49 localhost systemd[1]: tripleo_metrics_qdr.service: Main process exited, code=exited, status=139/n/a Nov 27 04:07:49 localhost systemd[1]: libpod-conmon-646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.scope: Deactivated successfully. Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.timer: Failed to open /run/systemd/transient/646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.timer: No such file or directory Nov 27 04:07:49 localhost systemd[1]: 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: Failed to open /run/systemd/transient/646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00.service: No such file or directory Nov 27 04:07:49 localhost podman[109863]: 2025-11-27 09:07:49.475119076 +0000 UTC m=+0.067509709 container cleanup 646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00 (image=registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1, name=metrics_qdr, architecture=x86_64, distribution-scope=public, config_id=tripleo_step1, name=rhosp17/openstack-qdrouterd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-qdrouterd, url=https://www.redhat.com, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'fcb31553ac96170fad0e8a9767bee027'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-qdrouterd:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'start_order': 1, 'user': 'qdrouterd', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/metrics_qdr.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/metrics_qdr:/var/lib/kolla/config_files/src:ro', '/var/lib/metrics_qdr:/var/lib/qdrouterd:z', '/var/log/containers/metrics_qdr:/var/log/qdrouterd:z']}, managed_by=tripleo_ansible, io.k8s.description=Red Hat OpenStack Platform 17.1 qdrouterd, release=1761123044, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 qdrouterd, io.openshift.expose-services=, io.buildah.version=1.41.4, build-date=2025-11-18T22:49:46Z, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=metrics_qdr, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 qdrouterd, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 qdrouterd, com.redhat.component=openstack-qdrouterd-container, vcs-ref=7ecaafae6fa9301c7dd5c0fca835eecf10dd147a, batch=17.1_20251118.1) Nov 27 04:07:49 localhost podman[109863]: metrics_qdr Nov 27 04:07:49 localhost systemd[1]: tripleo_metrics_qdr.service: Failed with result 'exit-code'. Nov 27 04:07:49 localhost systemd[1]: Stopped metrics_qdr container. Nov 27 04:07:50 localhost systemd[1]: var-lib-containers-storage-overlay-4e1c0acf5eef7e3d55db68ca8dcde6b2609fb9b33f28ab8549797d29f94ebb77-merged.mount: Deactivated successfully. Nov 27 04:07:50 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-646319e803d586a1616c7dadae098263d98bef728efc8756391a74a6929fac00-userdata-shm.mount: Deactivated successfully. Nov 27 04:07:50 localhost python3.9[109965]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_neutron_dhcp.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 4800.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:07:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25227 DF PROTO=TCP SPT=41390 DPT=9100 SEQ=3765640083 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C05DB0000000001030307) Nov 27 04:07:52 localhost python3.9[110058]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_neutron_l3_agent.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48305 DF PROTO=TCP SPT=50820 DPT=9101 SEQ=652401613 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C0DDA0000000001030307) Nov 27 04:07:53 localhost python3.9[110151]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_neutron_ovs_agent.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:54 localhost python3.9[110244]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_compute.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:07:55 localhost systemd[1]: Reloading. Nov 27 04:07:55 localhost systemd-rc-local-generator[110269]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:07:55 localhost systemd-sysv-generator[110274]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:07:55 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:07:55 localhost systemd[1]: Stopping nova_compute container... Nov 27 04:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 4800.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:07:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=26964 DF PROTO=TCP SPT=32802 DPT=9101 SEQ=2044993144 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C19DA0000000001030307) Nov 27 04:07:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48307 DF PROTO=TCP SPT=50820 DPT=9101 SEQ=652401613 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C259A0000000001030307) Nov 27 04:08:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13001 DF PROTO=TCP SPT=44890 DPT=9105 SEQ=3191363301 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C319A0000000001030307) Nov 27 04:08:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:08:06 localhost podman[110298]: 2025-11-27 09:08:06.751585847 +0000 UTC m=+0.087844467 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, maintainer=OpenStack TripleO Team, build-date=2025-11-19T00:36:58Z, config_id=tripleo_step4, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, vcs-type=git, release=1761123044, io.buildah.version=1.41.4, container_name=nova_migration_target, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64) Nov 27 04:08:07 localhost podman[110298]: 2025-11-27 09:08:07.149996957 +0000 UTC m=+0.486255567 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, config_id=tripleo_step4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, batch=17.1_20251118.1, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, release=1761123044, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, io.buildah.version=1.41.4, build-date=2025-11-19T00:36:58Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, name=rhosp17/openstack-nova-compute, vcs-type=git, container_name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:08:07 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:08:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48308 DF PROTO=TCP SPT=50820 DPT=9101 SEQ=652401613 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C45DA0000000001030307) Nov 27 04:08:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29191 DF PROTO=TCP SPT=43312 DPT=9882 SEQ=2676167498 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C48880000000001030307) Nov 27 04:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:08:08 localhost podman[110324]: Error: container ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 is not running Nov 27 04:08:08 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Main process exited, code=exited, status=125/n/a Nov 27 04:08:08 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed with result 'exit-code'. Nov 27 04:08:08 localhost podman[110322]: 2025-11-27 09:08:08.996873569 +0000 UTC m=+0.090173060 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, release=1761123044, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, tcib_managed=true, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, build-date=2025-11-19T00:14:25Z, io.openshift.expose-services=, version=17.1.12, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, container_name=ovn_metadata_agent, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 04:08:09 localhost systemd[1]: tmp-crun.5ugEJE.mount: Deactivated successfully. Nov 27 04:08:09 localhost podman[110323]: 2025-11-27 09:08:09.068761484 +0000 UTC m=+0.154505192 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, architecture=x86_64, config_id=tripleo_step4, com.redhat.component=openstack-ovn-controller-container, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, distribution-scope=public, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, url=https://www.redhat.com, batch=17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, vcs-type=git) Nov 27 04:08:09 localhost podman[110322]: 2025-11-27 09:08:09.086700808 +0000 UTC m=+0.180000289 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, distribution-scope=public, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, batch=17.1_20251118.1, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, config_id=tripleo_step4) Nov 27 04:08:09 localhost podman[110322]: unhealthy Nov 27 04:08:09 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:08:09 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:08:09 localhost podman[110323]: 2025-11-27 09:08:09.110908369 +0000 UTC m=+0.196652087 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, maintainer=OpenStack TripleO Team, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, vendor=Red Hat, Inc., org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, vcs-type=git, io.openshift.expose-services=, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, version=17.1.12, distribution-scope=public, batch=17.1_20251118.1, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, io.buildah.version=1.41.4) Nov 27 04:08:09 localhost podman[110323]: unhealthy Nov 27 04:08:09 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:08:09 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:08:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29193 DF PROTO=TCP SPT=43312 DPT=9882 SEQ=2676167498 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C549A0000000001030307) Nov 27 04:08:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17352 DF PROTO=TCP SPT=39334 DPT=9102 SEQ=181909759 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C63190000000001030307) Nov 27 04:08:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10602 DF PROTO=TCP SPT=45350 DPT=9100 SEQ=2912849491 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C6B5A0000000001030307) Nov 27 04:08:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10603 DF PROTO=TCP SPT=45350 DPT=9100 SEQ=2912849491 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C7B1A0000000001030307) Nov 27 04:08:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=34341 DF PROTO=TCP SPT=36432 DPT=9101 SEQ=3111451917 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C82DA0000000001030307) Nov 27 04:08:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18586 DF PROTO=TCP SPT=34832 DPT=9105 SEQ=2339646826 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C8F1A0000000001030307) Nov 27 04:08:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=34343 DF PROTO=TCP SPT=36432 DPT=9101 SEQ=3111451917 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127C9A9A0000000001030307) Nov 27 04:08:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18588 DF PROTO=TCP SPT=34832 DPT=9105 SEQ=2339646826 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CA6DA0000000001030307) Nov 27 04:08:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:08:37 localhost podman[110446]: 2025-11-27 09:08:37.508325794 +0000 UTC m=+0.072754161 container health_status f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, health_status=healthy, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, managed_by=tripleo_ansible, container_name=nova_migration_target, distribution-scope=public, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, build-date=2025-11-19T00:36:58Z, vcs-type=git, batch=17.1_20251118.1, config_id=tripleo_step4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, name=rhosp17/openstack-nova-compute, vendor=Red Hat, Inc., io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute) Nov 27 04:08:37 localhost podman[110446]: 2025-11-27 09:08:37.870379816 +0000 UTC m=+0.434808073 container exec_died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, container_name=nova_migration_target, url=https://www.redhat.com, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, name=rhosp17/openstack-nova-compute, tcib_managed=true, config_id=tripleo_step4, version=17.1.12, architecture=x86_64, build-date=2025-11-19T00:36:58Z, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, managed_by=tripleo_ansible, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.component=openstack-nova-compute-container, vendor=Red Hat, Inc., vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1) Nov 27 04:08:37 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Deactivated successfully. Nov 27 04:08:38 localhost podman[110285]: time="2025-11-27T09:08:38Z" level=warning msg="StopSignal SIGTERM failed to stop container nova_compute in 42 seconds, resorting to SIGKILL" Nov 27 04:08:38 localhost systemd[1]: session-c11.scope: Deactivated successfully. Nov 27 04:08:38 localhost systemd[1]: libpod-ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.scope: Deactivated successfully. Nov 27 04:08:38 localhost systemd[1]: libpod-ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.scope: Consumed 37.742s CPU time. Nov 27 04:08:38 localhost podman[110285]: 2025-11-27 09:08:38.084507102 +0000 UTC m=+42.094045382 container died ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-type=git, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, maintainer=OpenStack TripleO Team, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-nova-compute-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, container_name=nova_compute, vendor=Red Hat, Inc., build-date=2025-11-19T00:36:58Z, config_id=tripleo_step5, batch=17.1_20251118.1, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 nova-compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.timer: Deactivated successfully. Nov 27 04:08:38 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409. Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed to open /run/systemd/transient/ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: No such file or directory Nov 27 04:08:38 localhost podman[110285]: 2025-11-27 09:08:38.152037091 +0000 UTC m=+42.161575341 container cleanup ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_compute, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-compute, description=Red Hat OpenStack Platform 17.1 nova-compute, architecture=x86_64, maintainer=OpenStack TripleO Team, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, io.buildah.version=1.41.4, version=17.1.12, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step5, vcs-type=git, com.redhat.component=openstack-nova-compute-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, build-date=2025-11-19T00:36:58Z, name=rhosp17/openstack-nova-compute, batch=17.1_20251118.1) Nov 27 04:08:38 localhost podman[110285]: nova_compute Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.timer: Failed to open /run/systemd/transient/ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.timer: No such file or directory Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed to open /run/systemd/transient/ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: No such file or directory Nov 27 04:08:38 localhost podman[110471]: 2025-11-27 09:08:38.195371059 +0000 UTC m=+0.088611548 container cleanup ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, io.openshift.expose-services=, config_id=tripleo_step5, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, distribution-scope=public, batch=17.1_20251118.1, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, vendor=Red Hat, Inc., release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, managed_by=tripleo_ansible, container_name=nova_compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, tcib_managed=true, name=rhosp17/openstack-nova-compute, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-compute-container, build-date=2025-11-19T00:36:58Z, maintainer=OpenStack TripleO Team, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:08:38 localhost systemd[1]: libpod-conmon-ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.scope: Deactivated successfully. Nov 27 04:08:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=34344 DF PROTO=TCP SPT=36432 DPT=9101 SEQ=3111451917 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CBBDA0000000001030307) Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.timer: Failed to open /run/systemd/transient/ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.timer: No such file or directory Nov 27 04:08:38 localhost systemd[1]: ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: Failed to open /run/systemd/transient/ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409.service: No such file or directory Nov 27 04:08:38 localhost podman[110485]: 2025-11-27 09:08:38.29938632 +0000 UTC m=+0.072670338 container cleanup ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, tcib_managed=true, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, io.buildah.version=1.41.4, vcs-type=git, architecture=x86_64, maintainer=OpenStack TripleO Team, release=1761123044, summary=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_compute, name=rhosp17/openstack-nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, batch=17.1_20251118.1, url=https://www.redhat.com, config_id=tripleo_step5, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, vendor=Red Hat, Inc., distribution-scope=public, com.redhat.component=openstack-nova-compute-container) Nov 27 04:08:38 localhost podman[110485]: nova_compute Nov 27 04:08:38 localhost systemd[1]: tripleo_nova_compute.service: Deactivated successfully. Nov 27 04:08:38 localhost systemd[1]: Stopped nova_compute container. Nov 27 04:08:38 localhost systemd[1]: tripleo_nova_compute.service: Consumed 1.161s CPU time, no IO. Nov 27 04:08:38 localhost systemd[1]: var-lib-containers-storage-overlay-d183f9b9c8a709698b455a7b5d71cc9159fd44df5801ed4e9b97f3ae854b459f-merged.mount: Deactivated successfully. Nov 27 04:08:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60533 DF PROTO=TCP SPT=37184 DPT=9882 SEQ=1256864264 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CBDB60000000001030307) Nov 27 04:08:39 localhost python3.9[110589]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_migration_target.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:08:39 localhost systemd[1]: Reloading. Nov 27 04:08:39 localhost systemd-sysv-generator[110615]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:08:39 localhost systemd-rc-local-generator[110612]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:08:39 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:08:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:08:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:08:39 localhost systemd[1]: Stopping nova_migration_target container... Nov 27 04:08:39 localhost systemd[1]: libpod-f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.scope: Deactivated successfully. Nov 27 04:08:39 localhost systemd[1]: libpod-f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.scope: Consumed 35.047s CPU time. Nov 27 04:08:39 localhost podman[110641]: 2025-11-27 09:08:39.537163557 +0000 UTC m=+0.071965650 container died f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, build-date=2025-11-19T00:36:58Z, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_migration_target, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, release=1761123044, maintainer=OpenStack TripleO Team, version=17.1.12, io.buildah.version=1.41.4, batch=17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=rhosp17/openstack-nova-compute, vcs-type=git) Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.timer: Deactivated successfully. Nov 27 04:08:39 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33. Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Failed to open /run/systemd/transient/f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: No such file or directory Nov 27 04:08:39 localhost podman[110629]: 2025-11-27 09:08:39.603700578 +0000 UTC m=+0.146012254 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, summary=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, distribution-scope=public, com.redhat.component=openstack-ovn-controller-container, vcs-type=git, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, url=https://www.redhat.com, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, tcib_managed=true, maintainer=OpenStack TripleO Team, io.openshift.expose-services=, release=1761123044, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:08:39 localhost podman[110628]: 2025-11-27 09:08:39.651515646 +0000 UTC m=+0.193879313 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, vcs-type=git, config_id=tripleo_step4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, io.buildah.version=1.41.4, version=17.1.12, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, managed_by=tripleo_ansible, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, io.openshift.expose-services=, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent) Nov 27 04:08:39 localhost podman[110628]: 2025-11-27 09:08:39.669795218 +0000 UTC m=+0.212158935 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, release=1761123044, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.buildah.version=1.41.4, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn) Nov 27 04:08:39 localhost podman[110628]: unhealthy Nov 27 04:08:39 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:08:39 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:08:39 localhost podman[110641]: 2025-11-27 09:08:39.73856737 +0000 UTC m=+0.273369463 container cleanup f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, url=https://www.redhat.com, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-compute-container, description=Red Hat OpenStack Platform 17.1 nova-compute, build-date=2025-11-19T00:36:58Z, container_name=nova_migration_target, name=rhosp17/openstack-nova-compute, architecture=x86_64, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, vendor=Red Hat, Inc., release=1761123044, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, config_id=tripleo_step4, tcib_managed=true, io.buildah.version=1.41.4) Nov 27 04:08:39 localhost podman[110641]: nova_migration_target Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.timer: Failed to open /run/systemd/transient/f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.timer: No such file or directory Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Failed to open /run/systemd/transient/f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: No such file or directory Nov 27 04:08:39 localhost podman[110666]: 2025-11-27 09:08:39.757298935 +0000 UTC m=+0.211147247 container cleanup f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=openstack-nova-compute-container, container_name=nova_migration_target, tcib_managed=true, vendor=Red Hat, Inc., config_id=tripleo_step4, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 nova-compute, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, build-date=2025-11-19T00:36:58Z, vcs-type=git, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, description=Red Hat OpenStack Platform 17.1 nova-compute, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-compute) Nov 27 04:08:39 localhost podman[110629]: 2025-11-27 09:08:39.774365764 +0000 UTC m=+0.316677390 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, vcs-type=git, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., container_name=ovn_controller, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, build-date=2025-11-18T23:34:05Z, name=rhosp17/openstack-ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container) Nov 27 04:08:39 localhost systemd[1]: libpod-conmon-f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.scope: Deactivated successfully. Nov 27 04:08:39 localhost podman[110629]: unhealthy Nov 27 04:08:39 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:08:39 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.timer: Failed to open /run/systemd/transient/f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.timer: No such file or directory Nov 27 04:08:39 localhost systemd[1]: f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: Failed to open /run/systemd/transient/f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33.service: No such file or directory Nov 27 04:08:39 localhost podman[110694]: 2025-11-27 09:08:39.850158716 +0000 UTC m=+0.057729895 container cleanup f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_migration_target, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, url=https://www.redhat.com, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-compute, com.redhat.component=openstack-nova-compute-container, architecture=x86_64, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, vcs-type=git, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, container_name=nova_migration_target, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, distribution-scope=public, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/nova-migration-target.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/etc/ssh:/host-ssh:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared']}, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, tcib_managed=true, build-date=2025-11-19T00:36:58Z) Nov 27 04:08:39 localhost podman[110694]: nova_migration_target Nov 27 04:08:39 localhost systemd[1]: tripleo_nova_migration_target.service: Deactivated successfully. Nov 27 04:08:39 localhost systemd[1]: Stopped nova_migration_target container. Nov 27 04:08:40 localhost systemd[1]: tmp-crun.RP6RUf.mount: Deactivated successfully. Nov 27 04:08:40 localhost systemd[1]: var-lib-containers-storage-overlay-f46c77aa3acfba4abd8d96e809917692cad69270a1f593d1d2c9063645aac7e1-merged.mount: Deactivated successfully. Nov 27 04:08:40 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-f5298dfd9e34fbc235facffff6e6d7524ac852e0d5ce4f7ef9683ccc799cbc33-userdata-shm.mount: Deactivated successfully. Nov 27 04:08:40 localhost python3.9[110798]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtlogd_wrapper.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:08:40 localhost systemd[1]: Reloading. Nov 27 04:08:40 localhost systemd-sysv-generator[110827]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:08:40 localhost systemd-rc-local-generator[110822]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:08:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:08:41 localhost systemd[1]: Stopping nova_virtlogd_wrapper container... Nov 27 04:08:41 localhost systemd[1]: tmp-crun.ik4Qw8.mount: Deactivated successfully. Nov 27 04:08:41 localhost systemd[1]: libpod-c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458.scope: Deactivated successfully. Nov 27 04:08:41 localhost podman[110839]: 2025-11-27 09:08:41.119182114 +0000 UTC m=+0.085273937 container died c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, description=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step3, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, container_name=nova_virtlogd_wrapper, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, managed_by=tripleo_ansible, architecture=x86_64, build-date=2025-11-19T00:35:22Z, io.buildah.version=1.41.4, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, vcs-type=git, com.redhat.component=openstack-nova-libvirt-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, release=1761123044, version=17.1.12, vendor=Red Hat, Inc., konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:08:41 localhost podman[110839]: 2025-11-27 09:08:41.169017596 +0000 UTC m=+0.135109419 container cleanup c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, batch=17.1_20251118.1, distribution-scope=public, architecture=x86_64, build-date=2025-11-19T00:35:22Z, version=17.1.12, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, name=rhosp17/openstack-nova-libvirt, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, release=1761123044, config_id=tripleo_step3, vendor=Red Hat, Inc., container_name=nova_virtlogd_wrapper, url=https://www.redhat.com, description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-libvirt-container, io.openshift.expose-services=) Nov 27 04:08:41 localhost podman[110839]: nova_virtlogd_wrapper Nov 27 04:08:41 localhost podman[110852]: 2025-11-27 09:08:41.182669414 +0000 UTC m=+0.061397755 container cleanup c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, description=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, distribution-scope=public, managed_by=tripleo_ansible, container_name=nova_virtlogd_wrapper, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, tcib_managed=true, build-date=2025-11-19T00:35:22Z, architecture=x86_64, com.redhat.component=openstack-nova-libvirt-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, version=17.1.12, config_id=tripleo_step3) Nov 27 04:08:41 localhost systemd[1]: var-lib-containers-storage-overlay-2e54adbaf91d1dc2af01e4262ff4e93ee78842501d74a6c70e590105d0e0d8a6-merged.mount: Deactivated successfully. Nov 27 04:08:41 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458-userdata-shm.mount: Deactivated successfully. Nov 27 04:08:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60535 DF PROTO=TCP SPT=37184 DPT=9882 SEQ=1256864264 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CC9DB0000000001030307) Nov 27 04:08:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11214 DF PROTO=TCP SPT=37064 DPT=9102 SEQ=1112408081 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CD8490000000001030307) Nov 27 04:08:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64005 DF PROTO=TCP SPT=50706 DPT=9100 SEQ=4181575162 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CE05B0000000001030307) Nov 27 04:08:48 localhost systemd[1]: Stopping User Manager for UID 0... Nov 27 04:08:48 localhost systemd[85983]: Activating special unit Exit the Session... Nov 27 04:08:48 localhost systemd[85983]: Removed slice User Background Tasks Slice. Nov 27 04:08:48 localhost systemd[85983]: Stopped target Main User Target. Nov 27 04:08:48 localhost systemd[85983]: Stopped target Basic System. Nov 27 04:08:48 localhost systemd[85983]: Stopped target Paths. Nov 27 04:08:48 localhost systemd[85983]: Stopped target Sockets. Nov 27 04:08:48 localhost systemd[85983]: Stopped target Timers. Nov 27 04:08:48 localhost systemd[85983]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 04:08:48 localhost systemd[85983]: Closed D-Bus User Message Bus Socket. Nov 27 04:08:48 localhost systemd[85983]: Stopped Create User's Volatile Files and Directories. Nov 27 04:08:48 localhost systemd[85983]: Removed slice User Application Slice. Nov 27 04:08:48 localhost systemd[85983]: Reached target Shutdown. Nov 27 04:08:48 localhost systemd[85983]: Finished Exit the Session. Nov 27 04:08:48 localhost systemd[85983]: Reached target Exit the Session. Nov 27 04:08:48 localhost systemd[1]: user@0.service: Deactivated successfully. Nov 27 04:08:48 localhost systemd[1]: Stopped User Manager for UID 0. Nov 27 04:08:48 localhost systemd[1]: user@0.service: Consumed 5.305s CPU time, no IO. Nov 27 04:08:48 localhost systemd[1]: Stopping User Runtime Directory /run/user/0... Nov 27 04:08:48 localhost systemd[1]: run-user-0.mount: Deactivated successfully. Nov 27 04:08:48 localhost systemd[1]: user-runtime-dir@0.service: Deactivated successfully. Nov 27 04:08:48 localhost systemd[1]: Stopped User Runtime Directory /run/user/0. Nov 27 04:08:48 localhost systemd[1]: Removed slice User Slice of UID 0. Nov 27 04:08:48 localhost systemd[1]: user-0.slice: Consumed 6.267s CPU time. Nov 27 04:08:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64006 DF PROTO=TCP SPT=50706 DPT=9100 SEQ=4181575162 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CF01A0000000001030307) Nov 27 04:08:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25350 DF PROTO=TCP SPT=56804 DPT=9101 SEQ=3922976417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127CF81A0000000001030307) Nov 27 04:08:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48310 DF PROTO=TCP SPT=50820 DPT=9101 SEQ=652401613 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D03DA0000000001030307) Nov 27 04:08:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25352 DF PROTO=TCP SPT=56804 DPT=9101 SEQ=3922976417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D0FDB0000000001030307) Nov 27 04:09:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7430 DF PROTO=TCP SPT=45980 DPT=9105 SEQ=3677774210 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D1C1B0000000001030307) Nov 27 04:09:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25353 DF PROTO=TCP SPT=56804 DPT=9101 SEQ=3922976417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D2FDB0000000001030307) Nov 27 04:09:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35909 DF PROTO=TCP SPT=41810 DPT=9882 SEQ=3089084304 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D32E60000000001030307) Nov 27 04:09:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:09:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:09:10 localhost podman[110871]: 2025-11-27 09:09:10.043852968 +0000 UTC m=+0.111116623 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, distribution-scope=public, version=17.1.12, description=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vcs-type=git, release=1761123044, config_id=tripleo_step4, name=rhosp17/openstack-ovn-controller, architecture=x86_64, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, container_name=ovn_controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-18T23:34:05Z, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, io.buildah.version=1.41.4, tcib_managed=true) Nov 27 04:09:10 localhost podman[110871]: 2025-11-27 09:09:10.064991108 +0000 UTC m=+0.132254813 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, batch=17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, com.redhat.component=openstack-ovn-controller-container, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, architecture=x86_64, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, release=1761123044, container_name=ovn_controller, distribution-scope=public, io.openshift.expose-services=, io.buildah.version=1.41.4, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, maintainer=OpenStack TripleO Team) Nov 27 04:09:10 localhost podman[110871]: unhealthy Nov 27 04:09:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:09:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:09:10 localhost systemd[1]: tmp-crun.nan4pt.mount: Deactivated successfully. Nov 27 04:09:10 localhost podman[110870]: 2025-11-27 09:09:10.133951335 +0000 UTC m=+0.201032335 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.expose-services=, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, vendor=Red Hat, Inc., io.buildah.version=1.41.4, version=17.1.12, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, config_id=tripleo_step4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, container_name=ovn_metadata_agent, distribution-scope=public, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:09:10 localhost podman[110870]: 2025-11-27 09:09:10.1790409 +0000 UTC m=+0.246121880 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, url=https://www.redhat.com, release=1761123044, io.buildah.version=1.41.4, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:14:25Z, version=17.1.12, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, maintainer=OpenStack TripleO Team, vcs-type=git) Nov 27 04:09:10 localhost podman[110870]: unhealthy Nov 27 04:09:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:09:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:09:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35911 DF PROTO=TCP SPT=41810 DPT=9882 SEQ=3089084304 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D3EDA0000000001030307) Nov 27 04:09:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58967 DF PROTO=TCP SPT=47720 DPT=9102 SEQ=2721227534 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D4D790000000001030307) Nov 27 04:09:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18004 DF PROTO=TCP SPT=58404 DPT=9100 SEQ=3543579280 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D559A0000000001030307) Nov 27 04:09:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18005 DF PROTO=TCP SPT=58404 DPT=9100 SEQ=3543579280 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D655A0000000001030307) Nov 27 04:09:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10265 DF PROTO=TCP SPT=56874 DPT=9101 SEQ=2290328222 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D6D5B0000000001030307) Nov 27 04:09:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32718 DF PROTO=TCP SPT=37802 DPT=9105 SEQ=1375972534 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D795A0000000001030307) Nov 27 04:09:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10267 DF PROTO=TCP SPT=56874 DPT=9101 SEQ=2290328222 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D851A0000000001030307) Nov 27 04:09:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32720 DF PROTO=TCP SPT=37802 DPT=9105 SEQ=1375972534 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127D911B0000000001030307) Nov 27 04:09:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10268 DF PROTO=TCP SPT=56874 DPT=9101 SEQ=2290328222 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DA5DA0000000001030307) Nov 27 04:09:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28397 DF PROTO=TCP SPT=44824 DPT=9882 SEQ=3848798372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DA8160000000001030307) Nov 27 04:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:09:40 localhost systemd[1]: tmp-crun.c1I9TL.mount: Deactivated successfully. Nov 27 04:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:09:40 localhost podman[110987]: 2025-11-27 09:09:40.268764573 +0000 UTC m=+0.107051395 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., release=1761123044, version=17.1.12, container_name=ovn_controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, managed_by=tripleo_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, build-date=2025-11-18T23:34:05Z, config_id=tripleo_step4, maintainer=OpenStack TripleO Team, tcib_managed=true, com.redhat.component=openstack-ovn-controller-container, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:09:40 localhost podman[110987]: 2025-11-27 09:09:40.313013834 +0000 UTC m=+0.151300646 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.buildah.version=1.41.4, container_name=ovn_controller, architecture=x86_64, distribution-scope=public, config_id=tripleo_step4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, version=17.1.12, url=https://www.redhat.com, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:09:40 localhost podman[110987]: unhealthy Nov 27 04:09:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:09:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:09:40 localhost podman[111003]: 2025-11-27 09:09:40.368476898 +0000 UTC m=+0.090131788 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.expose-services=, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, architecture=x86_64, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_metadata_agent, release=1761123044, name=rhosp17/openstack-neutron-metadata-agent-ovn, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-type=git, build-date=2025-11-19T00:14:25Z, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05) Nov 27 04:09:40 localhost podman[111003]: 2025-11-27 09:09:40.410951682 +0000 UTC m=+0.132606632 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, tcib_managed=true, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, release=1761123044, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, container_name=ovn_metadata_agent, url=https://www.redhat.com, build-date=2025-11-19T00:14:25Z, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, vcs-type=git, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-neutron-metadata-agent-ovn, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, vendor=Red Hat, Inc., io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:09:40 localhost podman[111003]: unhealthy Nov 27 04:09:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:09:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:09:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28399 DF PROTO=TCP SPT=44824 DPT=9882 SEQ=3848798372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DB41A0000000001030307) Nov 27 04:09:41 localhost systemd[1]: Starting Check and recover tripleo_nova_virtqemud... Nov 27 04:09:41 localhost recover_tripleo_nova_virtqemud[111029]: 63639 Nov 27 04:09:41 localhost systemd[1]: tripleo_nova_virtqemud_recover.service: Deactivated successfully. Nov 27 04:09:41 localhost systemd[1]: Finished Check and recover tripleo_nova_virtqemud. Nov 27 04:09:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60981 DF PROTO=TCP SPT=51804 DPT=9102 SEQ=915979109 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DC2AA0000000001030307) Nov 27 04:09:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41526 DF PROTO=TCP SPT=42852 DPT=9100 SEQ=3776113182 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DCADB0000000001030307) Nov 27 04:09:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41527 DF PROTO=TCP SPT=42852 DPT=9100 SEQ=3776113182 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DDA9A0000000001030307) Nov 27 04:09:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9842 DF PROTO=TCP SPT=55742 DPT=9101 SEQ=2995578484 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DE29A0000000001030307) Nov 27 04:09:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25355 DF PROTO=TCP SPT=56804 DPT=9101 SEQ=3922976417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DEDDB0000000001030307) Nov 27 04:09:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7433 DF PROTO=TCP SPT=45980 DPT=9105 SEQ=3677774210 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127DF9DA0000000001030307) Nov 27 04:10:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44638 DF PROTO=TCP SPT=57594 DPT=9105 SEQ=2788622992 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E065A0000000001030307) Nov 27 04:10:05 localhost systemd[1]: tripleo_nova_virtlogd_wrapper.service: State 'stop-sigterm' timed out. Killing. Nov 27 04:10:05 localhost systemd[1]: tripleo_nova_virtlogd_wrapper.service: Killing process 62863 (conmon) with signal SIGKILL. Nov 27 04:10:05 localhost systemd[1]: tripleo_nova_virtlogd_wrapper.service: Main process exited, code=killed, status=9/KILL Nov 27 04:10:05 localhost systemd[1]: libpod-conmon-c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458.scope: Deactivated successfully. Nov 27 04:10:05 localhost podman[111043]: error opening file `/run/crun/c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458/status`: No such file or directory Nov 27 04:10:05 localhost systemd[1]: tmp-crun.5KHY2h.mount: Deactivated successfully. Nov 27 04:10:05 localhost podman[111031]: 2025-11-27 09:10:05.489297216 +0000 UTC m=+0.082174004 container cleanup c335e4ca13180dedb22d05f17cb5a0ed25c5632e26f124886395bc99542bf458 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd_wrapper, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, container_name=nova_virtlogd_wrapper, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, vendor=Red Hat, Inc., version=17.1.12, com.redhat.component=openstack-nova-libvirt-container, build-date=2025-11-19T00:35:22Z, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 0, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtlogd.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/container-config-scripts/virtlogd_wrapper:/usr/local/bin/virtlogd_wrapper:ro']}, config_id=tripleo_step3, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, release=1761123044, name=rhosp17/openstack-nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64) Nov 27 04:10:05 localhost podman[111031]: nova_virtlogd_wrapper Nov 27 04:10:05 localhost systemd[1]: tmp-crun.38OvWF.mount: Deactivated successfully. Nov 27 04:10:05 localhost systemd[1]: tripleo_nova_virtlogd_wrapper.service: Failed with result 'timeout'. Nov 27 04:10:05 localhost systemd[1]: Stopped nova_virtlogd_wrapper container. Nov 27 04:10:06 localhost python3.9[111136]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtnodedevd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:10:06 localhost systemd[1]: Reloading. Nov 27 04:10:06 localhost systemd-sysv-generator[111164]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:10:06 localhost systemd-rc-local-generator[111159]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:10:06 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:10:06 localhost systemd[1]: Stopping nova_virtnodedevd container... Nov 27 04:10:06 localhost systemd[1]: libpod-bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22.scope: Deactivated successfully. Nov 27 04:10:06 localhost systemd[1]: libpod-bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22.scope: Consumed 1.633s CPU time. Nov 27 04:10:06 localhost podman[111176]: 2025-11-27 09:10:06.686917231 +0000 UTC m=+0.074949860 container died bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, architecture=x86_64, distribution-scope=public, vendor=Red Hat, Inc., config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=nova_virtnodedevd, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.component=openstack-nova-libvirt-container, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, maintainer=OpenStack TripleO Team, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, version=17.1.12, release=1761123044) Nov 27 04:10:06 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22-userdata-shm.mount: Deactivated successfully. Nov 27 04:10:06 localhost podman[111176]: 2025-11-27 09:10:06.725340736 +0000 UTC m=+0.113373305 container cleanup bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, config_id=tripleo_step3, batch=17.1_20251118.1, io.buildah.version=1.41.4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, tcib_managed=true, version=17.1.12, architecture=x86_64, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, url=https://www.redhat.com, container_name=nova_virtnodedevd, com.redhat.component=openstack-nova-libvirt-container, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, name=rhosp17/openstack-nova-libvirt, distribution-scope=public, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, io.openshift.expose-services=, build-date=2025-11-19T00:35:22Z) Nov 27 04:10:06 localhost podman[111176]: nova_virtnodedevd Nov 27 04:10:06 localhost podman[111190]: 2025-11-27 09:10:06.771110288 +0000 UTC m=+0.070794047 container cleanup bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, vendor=Red Hat, Inc., io.openshift.expose-services=, name=rhosp17/openstack-nova-libvirt, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, vcs-type=git, tcib_managed=true, url=https://www.redhat.com, build-date=2025-11-19T00:35:22Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, io.buildah.version=1.41.4, version=17.1.12, release=1761123044, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=nova_virtnodedevd, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, distribution-scope=public, com.redhat.component=openstack-nova-libvirt-container) Nov 27 04:10:06 localhost systemd[1]: libpod-conmon-bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22.scope: Deactivated successfully. Nov 27 04:10:06 localhost podman[111220]: error opening file `/run/crun/bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22/status`: No such file or directory Nov 27 04:10:06 localhost podman[111206]: 2025-11-27 09:10:06.905283512 +0000 UTC m=+0.097430145 container cleanup bc3f56ca5fdaee48c743e8c7e3d2536524a86cd48073248482652adaf9be6b22 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtnodedevd, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-nova-libvirt-container, url=https://www.redhat.com, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 2, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtnodedevd.json:/var/lib/kolla/config_files/config.json:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, io.openshift.expose-services=, version=17.1.12, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:35:22Z, maintainer=OpenStack TripleO Team, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-libvirt, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_virtnodedevd, release=1761123044, architecture=x86_64, tcib_managed=true, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step3, managed_by=tripleo_ansible) Nov 27 04:10:06 localhost podman[111206]: nova_virtnodedevd Nov 27 04:10:06 localhost systemd[1]: tripleo_nova_virtnodedevd.service: Deactivated successfully. Nov 27 04:10:06 localhost systemd[1]: Stopped nova_virtnodedevd container. Nov 27 04:10:07 localhost python3.9[111313]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtproxyd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:10:07 localhost systemd[1]: Reloading. Nov 27 04:10:07 localhost systemd-rc-local-generator[111338]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:10:07 localhost systemd-sysv-generator[111342]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:10:07 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:10:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9845 DF PROTO=TCP SPT=55742 DPT=9101 SEQ=2995578484 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E19DA0000000001030307) Nov 27 04:10:07 localhost systemd[1]: var-lib-containers-storage-overlay-45273e7102bd399d5b266d522d0071b485c00ef9205c7542b9597c3a41f80098-merged.mount: Deactivated successfully. Nov 27 04:10:07 localhost systemd[1]: Stopping nova_virtproxyd container... Nov 27 04:10:08 localhost systemd[1]: tmp-crun.HiVQtk.mount: Deactivated successfully. Nov 27 04:10:08 localhost systemd[1]: libpod-6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172.scope: Deactivated successfully. Nov 27 04:10:08 localhost podman[111354]: 2025-11-27 09:10:08.042253874 +0000 UTC m=+0.082432972 container died 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, name=rhosp17/openstack-nova-libvirt, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, maintainer=OpenStack TripleO Team, vcs-type=git, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=nova_virtproxyd, version=17.1.12, config_id=tripleo_step3, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, tcib_managed=true, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, com.redhat.component=openstack-nova-libvirt-container, description=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:10:08 localhost systemd[1]: tmp-crun.Jdv1ty.mount: Deactivated successfully. Nov 27 04:10:08 localhost podman[111354]: 2025-11-27 09:10:08.104280984 +0000 UTC m=+0.144460052 container cleanup 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, description=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-libvirt-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, batch=17.1_20251118.1, build-date=2025-11-19T00:35:22Z, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, managed_by=tripleo_ansible, name=rhosp17/openstack-nova-libvirt, container_name=nova_virtproxyd, io.openshift.expose-services=, vcs-type=git, config_id=tripleo_step3, release=1761123044, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4) Nov 27 04:10:08 localhost podman[111354]: nova_virtproxyd Nov 27 04:10:08 localhost podman[111368]: 2025-11-27 09:10:08.178974696 +0000 UTC m=+0.126620791 container cleanup 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, architecture=x86_64, io.openshift.expose-services=, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, release=1761123044, maintainer=OpenStack TripleO Team, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, container_name=nova_virtproxyd, com.redhat.component=openstack-nova-libvirt-container, io.buildah.version=1.41.4, config_id=tripleo_step3, batch=17.1_20251118.1, build-date=2025-11-19T00:35:22Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, distribution-scope=public, name=rhosp17/openstack-nova-libvirt) Nov 27 04:10:08 localhost systemd[1]: libpod-conmon-6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172.scope: Deactivated successfully. Nov 27 04:10:08 localhost podman[111397]: error opening file `/run/crun/6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172/status`: No such file or directory Nov 27 04:10:08 localhost podman[111386]: 2025-11-27 09:10:08.276798281 +0000 UTC m=+0.066038600 container cleanup 6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtproxyd, com.redhat.component=openstack-nova-libvirt-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, io.openshift.expose-services=, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, version=17.1.12, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_virtproxyd, batch=17.1_20251118.1, io.buildah.version=1.41.4, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, config_id=tripleo_step3, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 5, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtproxyd.json:/var/lib/kolla/config_files/config.json:ro']}, build-date=2025-11-19T00:35:22Z, release=1761123044) Nov 27 04:10:08 localhost podman[111386]: nova_virtproxyd Nov 27 04:10:08 localhost systemd[1]: tripleo_nova_virtproxyd.service: Deactivated successfully. Nov 27 04:10:08 localhost systemd[1]: Stopped nova_virtproxyd container. Nov 27 04:10:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36854 DF PROTO=TCP SPT=35292 DPT=9882 SEQ=3015373684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E1D470000000001030307) Nov 27 04:10:08 localhost systemd[1]: var-lib-containers-storage-overlay-8755cddd819f8de1c9380e57820b4fe24d56b0250981fcd203a32a1f1bd38dff-merged.mount: Deactivated successfully. Nov 27 04:10:08 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-6f043730c740799c6479e4fc0aea7478b8a98eaf4e65a2d632ee9922c0dbb172-userdata-shm.mount: Deactivated successfully. Nov 27 04:10:09 localhost python3.9[111490]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtqemud.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:10:10 localhost systemd[1]: Reloading. Nov 27 04:10:10 localhost systemd-sysv-generator[111522]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:10:10 localhost systemd-rc-local-generator[111517]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:10:10 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:10:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:10:10 localhost systemd[1]: tripleo_nova_virtqemud_recover.timer: Deactivated successfully. Nov 27 04:10:10 localhost systemd[1]: Stopped Check and recover tripleo_nova_virtqemud every 10m. Nov 27 04:10:10 localhost systemd[1]: Stopping nova_virtqemud container... Nov 27 04:10:10 localhost systemd[1]: tmp-crun.4diJgv.mount: Deactivated successfully. Nov 27 04:10:10 localhost podman[111532]: 2025-11-27 09:10:10.493810591 +0000 UTC m=+0.081608659 container died c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, io.openshift.expose-services=, version=17.1.12, com.redhat.component=openstack-nova-libvirt-container, config_id=tripleo_step3, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, tcib_managed=true, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, konflux.additional-tags=17.1.12 17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, container_name=nova_virtqemud, vendor=Red Hat, Inc., vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, release=1761123044, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt) Nov 27 04:10:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:10:10 localhost systemd[1]: libpod-c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb.scope: Deactivated successfully. Nov 27 04:10:10 localhost systemd[1]: libpod-c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb.scope: Consumed 2.952s CPU time. Nov 27 04:10:10 localhost podman[111530]: 2025-11-27 09:10:10.535456913 +0000 UTC m=+0.133331733 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, maintainer=OpenStack TripleO Team, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.buildah.version=1.41.4, batch=17.1_20251118.1, release=1761123044, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, url=https://www.redhat.com, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, distribution-scope=public, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, tcib_managed=true, container_name=ovn_controller, io.openshift.expose-services=, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=17.1.12, name=rhosp17/openstack-ovn-controller, build-date=2025-11-18T23:34:05Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vendor=Red Hat, Inc.) Nov 27 04:10:10 localhost podman[111530]: 2025-11-27 09:10:10.55279797 +0000 UTC m=+0.150672750 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, release=1761123044, io.openshift.expose-services=, tcib_managed=true, version=17.1.12, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.component=openstack-ovn-controller-container, summary=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, name=rhosp17/openstack-ovn-controller, io.buildah.version=1.41.4, config_id=tripleo_step4, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, vendor=Red Hat, Inc.) Nov 27 04:10:10 localhost podman[111530]: unhealthy Nov 27 04:10:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:10:10 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:10:10 localhost podman[111532]: 2025-11-27 09:10:10.573299921 +0000 UTC m=+0.161097979 container cleanup c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, konflux.additional-tags=17.1.12 17.1_20251118.1, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, name=rhosp17/openstack-nova-libvirt, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, maintainer=OpenStack TripleO Team, config_id=tripleo_step3, batch=17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, io.buildah.version=1.41.4, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, managed_by=tripleo_ansible, vendor=Red Hat, Inc., release=1761123044, container_name=nova_virtqemud, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:10:10 localhost podman[111532]: nova_virtqemud Nov 27 04:10:10 localhost podman[111558]: 2025-11-27 09:10:10.644812388 +0000 UTC m=+0.132797547 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.openshift.expose-services=, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, io.buildah.version=1.41.4, vendor=Red Hat, Inc., config_id=tripleo_step4, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, name=rhosp17/openstack-neutron-metadata-agent-ovn, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, build-date=2025-11-19T00:14:25Z, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, tcib_managed=true, release=1761123044) Nov 27 04:10:10 localhost podman[111558]: 2025-11-27 09:10:10.658362083 +0000 UTC m=+0.146347222 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, build-date=2025-11-19T00:14:25Z, url=https://www.redhat.com, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, managed_by=tripleo_ansible, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, config_id=tripleo_step4, vcs-type=git, tcib_managed=true, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, io.openshift.expose-services=, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, name=rhosp17/openstack-neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, batch=17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:10:10 localhost podman[111558]: unhealthy Nov 27 04:10:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:10:10 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:10:10 localhost podman[111557]: 2025-11-27 09:10:10.71245648 +0000 UTC m=+0.212010612 container cleanup c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, vcs-type=git, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, io.openshift.expose-services=, tcib_managed=true, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, distribution-scope=public, release=1761123044, url=https://www.redhat.com, name=rhosp17/openstack-nova-libvirt, version=17.1.12, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, container_name=nova_virtqemud, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-libvirt) Nov 27 04:10:11 localhost systemd[1]: var-lib-containers-storage-overlay-d966a5434bed697fddf61042b9322213ec3395b654bd0213cb1773bf539000a2-merged.mount: Deactivated successfully. Nov 27 04:10:11 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb-userdata-shm.mount: Deactivated successfully. Nov 27 04:10:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36856 DF PROTO=TCP SPT=35292 DPT=9882 SEQ=3015373684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E295B0000000001030307) Nov 27 04:10:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35832 DF PROTO=TCP SPT=36564 DPT=9102 SEQ=3594029527 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E37D80000000001030307) Nov 27 04:10:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6675 DF PROTO=TCP SPT=59338 DPT=9100 SEQ=584560227 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E401A0000000001030307) Nov 27 04:10:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6676 DF PROTO=TCP SPT=59338 DPT=9100 SEQ=584560227 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E4FDA0000000001030307) Nov 27 04:10:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58536 DF PROTO=TCP SPT=59808 DPT=9101 SEQ=1415272487 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E579A0000000001030307) Nov 27 04:10:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10270 DF PROTO=TCP SPT=56874 DPT=9101 SEQ=2290328222 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E63DA0000000001030307) Nov 27 04:10:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58538 DF PROTO=TCP SPT=59808 DPT=9101 SEQ=1415272487 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E6F5A0000000001030307) Nov 27 04:10:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=30678 DF PROTO=TCP SPT=60486 DPT=9105 SEQ=840460466 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E7B9A0000000001030307) Nov 27 04:10:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58539 DF PROTO=TCP SPT=59808 DPT=9101 SEQ=1415272487 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E8FDB0000000001030307) Nov 27 04:10:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38536 DF PROTO=TCP SPT=35088 DPT=9882 SEQ=848671845 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E92770000000001030307) Nov 27 04:10:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:10:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:10:40 localhost podman[111729]: 2025-11-27 09:10:40.757184261 +0000 UTC m=+0.101551036 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, architecture=x86_64, io.openshift.expose-services=, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, io.buildah.version=1.41.4, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 ovn-controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, container_name=ovn_controller, maintainer=OpenStack TripleO Team, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, distribution-scope=public, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., batch=17.1_20251118.1, config_id=tripleo_step4, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, version=17.1.12) Nov 27 04:10:40 localhost podman[111729]: 2025-11-27 09:10:40.77498309 +0000 UTC m=+0.119349865 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, url=https://www.redhat.com, batch=17.1_20251118.1, managed_by=tripleo_ansible, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, io.buildah.version=1.41.4, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, vendor=Red Hat, Inc., version=17.1.12, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, build-date=2025-11-18T23:34:05Z, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller) Nov 27 04:10:40 localhost podman[111729]: unhealthy Nov 27 04:10:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:10:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:10:40 localhost podman[111745]: 2025-11-27 09:10:40.845808989 +0000 UTC m=+0.078320021 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, distribution-scope=public, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.expose-services=, io.buildah.version=1.41.4, vendor=Red Hat, Inc., release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:10:40 localhost podman[111745]: 2025-11-27 09:10:40.859168248 +0000 UTC m=+0.091679270 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_id=tripleo_step4, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, version=17.1.12, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, architecture=x86_64, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, managed_by=tripleo_ansible, batch=17.1_20251118.1, io.openshift.expose-services=, vendor=Red Hat, Inc., org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, vcs-type=git, distribution-scope=public, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container) Nov 27 04:10:40 localhost podman[111745]: unhealthy Nov 27 04:10:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:10:40 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:10:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38538 DF PROTO=TCP SPT=35088 DPT=9882 SEQ=848671845 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127E9E9A0000000001030307) Nov 27 04:10:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=5734 DF PROTO=TCP SPT=38866 DPT=9102 SEQ=357472770 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127EAD090000000001030307) Nov 27 04:10:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55433 DF PROTO=TCP SPT=60796 DPT=9100 SEQ=2578279207 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127EB51A0000000001030307) Nov 27 04:10:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55434 DF PROTO=TCP SPT=60796 DPT=9100 SEQ=2578279207 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127EC4DA0000000001030307) Nov 27 04:10:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12030 DF PROTO=TCP SPT=40594 DPT=9101 SEQ=1539121357 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127ECCDA0000000001030307) Nov 27 04:10:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18427 DF PROTO=TCP SPT=42546 DPT=9105 SEQ=1686544473 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127ED91A0000000001030307) Nov 27 04:10:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44641 DF PROTO=TCP SPT=57594 DPT=9105 SEQ=2788622992 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127EE3DA0000000001030307) Nov 27 04:11:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18429 DF PROTO=TCP SPT=42546 DPT=9105 SEQ=1686544473 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127EF0DA0000000001030307) Nov 27 04:11:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12033 DF PROTO=TCP SPT=40594 DPT=9101 SEQ=1539121357 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F05DA0000000001030307) Nov 27 04:11:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20060 DF PROTO=TCP SPT=42362 DPT=9882 SEQ=720359462 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F07A90000000001030307) Nov 27 04:11:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:11:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:11:10 localhost podman[111767]: 2025-11-27 09:11:10.990641165 +0000 UTC m=+0.084552868 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, container_name=ovn_metadata_agent, tcib_managed=true, architecture=x86_64, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, vcs-type=git, release=1761123044, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c) Nov 27 04:11:11 localhost podman[111767]: 2025-11-27 09:11:11.00493987 +0000 UTC m=+0.098851613 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, vcs-type=git, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, batch=17.1_20251118.1, distribution-scope=public, config_id=tripleo_step4, build-date=2025-11-19T00:14:25Z, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=tripleo_ansible, architecture=x86_64, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, tcib_managed=true, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}) Nov 27 04:11:11 localhost podman[111767]: unhealthy Nov 27 04:11:11 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:11:11 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:11:11 localhost podman[111768]: 2025-11-27 09:11:11.088266834 +0000 UTC m=+0.180027769 container health_status aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, health_status=unhealthy, build-date=2025-11-18T23:34:05Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, architecture=x86_64, vendor=Red Hat, Inc., maintainer=OpenStack TripleO Team, url=https://www.redhat.com, release=1761123044, container_name=ovn_controller, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, config_id=tripleo_step4, managed_by=tripleo_ansible, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, distribution-scope=public) Nov 27 04:11:11 localhost podman[111768]: 2025-11-27 09:11:11.131969032 +0000 UTC m=+0.223730027 container exec_died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 ovn-controller, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, container_name=ovn_controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, batch=17.1_20251118.1, build-date=2025-11-18T23:34:05Z, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., url=https://www.redhat.com, io.buildah.version=1.41.4, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, release=1761123044, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, managed_by=tripleo_ansible, name=rhosp17/openstack-ovn-controller) Nov 27 04:11:11 localhost podman[111768]: unhealthy Nov 27 04:11:11 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:11:11 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed with result 'exit-code'. Nov 27 04:11:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20062 DF PROTO=TCP SPT=42362 DPT=9882 SEQ=720359462 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F139B0000000001030307) Nov 27 04:11:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19184 DF PROTO=TCP SPT=58438 DPT=9102 SEQ=3936158115 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F22390000000001030307) Nov 27 04:11:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54947 DF PROTO=TCP SPT=36896 DPT=9100 SEQ=3712107319 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F2A5B0000000001030307) Nov 27 04:11:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54948 DF PROTO=TCP SPT=36896 DPT=9100 SEQ=3712107319 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F3A1A0000000001030307) Nov 27 04:11:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=65073 DF PROTO=TCP SPT=58638 DPT=9101 SEQ=2504457277 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F421B0000000001030307) Nov 27 04:11:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58541 DF PROTO=TCP SPT=59808 DPT=9101 SEQ=1415272487 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F4DDB0000000001030307) Nov 27 04:11:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=65075 DF PROTO=TCP SPT=58638 DPT=9101 SEQ=2504457277 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F59DA0000000001030307) Nov 27 04:11:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31332 DF PROTO=TCP SPT=52230 DPT=9105 SEQ=969979810 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F65DA0000000001030307) Nov 27 04:11:34 localhost systemd[1]: tripleo_nova_virtqemud.service: State 'stop-sigterm' timed out. Killing. Nov 27 04:11:34 localhost systemd[1]: tripleo_nova_virtqemud.service: Killing process 63635 (conmon) with signal SIGKILL. Nov 27 04:11:34 localhost systemd[1]: tripleo_nova_virtqemud.service: Main process exited, code=killed, status=9/KILL Nov 27 04:11:34 localhost systemd[1]: libpod-conmon-c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb.scope: Deactivated successfully. Nov 27 04:11:34 localhost systemd[1]: tmp-crun.s7mCMa.mount: Deactivated successfully. Nov 27 04:11:34 localhost podman[111895]: error opening file `/run/crun/c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb/status`: No such file or directory Nov 27 04:11:34 localhost podman[111884]: 2025-11-27 09:11:34.741818394 +0000 UTC m=+0.077388282 container cleanup c43af16cc2838d46414dea3861a5303fcb27b2928cba64509c699497c3e95aeb (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtqemud, distribution-scope=public, name=rhosp17/openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, version=17.1.12, io.openshift.expose-services=, container_name=nova_virtqemud, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, vcs-type=git, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 4, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtqemud.json:/var/lib/kolla/config_files/config.json:ro', '/var/log/containers/libvirt/swtpm:/var/log/swtpm:z']}, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, release=1761123044, architecture=x86_64, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:11:34 localhost podman[111884]: nova_virtqemud Nov 27 04:11:34 localhost systemd[1]: tripleo_nova_virtqemud.service: Failed with result 'timeout'. Nov 27 04:11:34 localhost systemd[1]: Stopped nova_virtqemud container. Nov 27 04:11:35 localhost python3.9[111988]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtqemud_recover.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:11:35 localhost systemd[1]: Reloading. Nov 27 04:11:35 localhost systemd-sysv-generator[112019]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:11:35 localhost systemd-rc-local-generator[112015]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:11:35 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:11:36 localhost python3.9[112117]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtsecretd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:11:37 localhost systemd[1]: Reloading. Nov 27 04:11:37 localhost systemd-sysv-generator[112147]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:11:37 localhost systemd-rc-local-generator[112141]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:11:37 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:11:37 localhost systemd[1]: Stopping nova_virtsecretd container... Nov 27 04:11:37 localhost systemd[1]: libpod-894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c.scope: Deactivated successfully. Nov 27 04:11:37 localhost podman[112157]: 2025-11-27 09:11:37.927030294 +0000 UTC m=+0.077618668 container died 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, vendor=Red Hat, Inc., distribution-scope=public, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, name=rhosp17/openstack-nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, tcib_managed=true, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, version=17.1.12, batch=17.1_20251118.1, container_name=nova_virtsecretd, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, managed_by=tripleo_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step3, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, url=https://www.redhat.com) Nov 27 04:11:37 localhost podman[112157]: 2025-11-27 09:11:37.980022889 +0000 UTC m=+0.130611273 container cleanup 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-type=git, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, konflux.additional-tags=17.1.12 17.1_20251118.1, distribution-scope=public, maintainer=OpenStack TripleO Team, release=1761123044, name=rhosp17/openstack-nova-libvirt, batch=17.1_20251118.1, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, com.redhat.component=openstack-nova-libvirt-container, container_name=nova_virtsecretd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vendor=Red Hat, Inc., architecture=x86_64, managed_by=tripleo_ansible, config_id=tripleo_step3, tcib_managed=true, url=https://www.redhat.com, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:11:37 localhost podman[112157]: nova_virtsecretd Nov 27 04:11:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=65076 DF PROTO=TCP SPT=58638 DPT=9101 SEQ=2504457277 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F79DA0000000001030307) Nov 27 04:11:38 localhost podman[112171]: 2025-11-27 09:11:38.014094695 +0000 UTC m=+0.077040873 container cleanup 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, build-date=2025-11-19T00:35:22Z, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1, managed_by=tripleo_ansible, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, version=17.1.12, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, distribution-scope=public, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, description=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step3, vcs-type=git, com.redhat.component=openstack-nova-libvirt-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, url=https://www.redhat.com, io.buildah.version=1.41.4, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, container_name=nova_virtsecretd, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044) Nov 27 04:11:38 localhost systemd[1]: libpod-conmon-894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c.scope: Deactivated successfully. Nov 27 04:11:38 localhost podman[112198]: error opening file `/run/crun/894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c/status`: No such file or directory Nov 27 04:11:38 localhost podman[112186]: 2025-11-27 09:11:38.122125849 +0000 UTC m=+0.069256232 container cleanup 894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtsecretd, io.openshift.expose-services=, config_id=tripleo_step3, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, distribution-scope=public, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, url=https://www.redhat.com, tcib_managed=true, build-date=2025-11-19T00:35:22Z, batch=17.1_20251118.1, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, maintainer=OpenStack TripleO Team, container_name=nova_virtsecretd, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 1, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtsecretd.json:/var/lib/kolla/config_files/config.json:ro']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, name=rhosp17/openstack-nova-libvirt, io.buildah.version=1.41.4, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vendor=Red Hat, Inc., com.redhat.component=openstack-nova-libvirt-container) Nov 27 04:11:38 localhost podman[112186]: nova_virtsecretd Nov 27 04:11:38 localhost systemd[1]: tripleo_nova_virtsecretd.service: Deactivated successfully. Nov 27 04:11:38 localhost systemd[1]: Stopped nova_virtsecretd container. Nov 27 04:11:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36076 DF PROTO=TCP SPT=56752 DPT=9882 SEQ=739974169 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F7CD70000000001030307) Nov 27 04:11:38 localhost systemd[1]: tmp-crun.hWVZBV.mount: Deactivated successfully. Nov 27 04:11:38 localhost systemd[1]: var-lib-containers-storage-overlay-eb0cba9abbd60c28b3923769313771c6cd1bc7a3f12cb9fdd423d5ab00f00aac-merged.mount: Deactivated successfully. Nov 27 04:11:38 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-894398dc5cebb0d392c460b20e5ee913efad9011a15ce31b6bd155e9f72c700c-userdata-shm.mount: Deactivated successfully. Nov 27 04:11:38 localhost python3.9[112291]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtstoraged.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:11:39 localhost systemd[1]: Reloading. Nov 27 04:11:39 localhost systemd-rc-local-generator[112317]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:11:39 localhost systemd-sysv-generator[112323]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:11:39 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:11:39 localhost systemd[1]: Stopping nova_virtstoraged container... Nov 27 04:11:39 localhost systemd[1]: libpod-c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4.scope: Deactivated successfully. Nov 27 04:11:39 localhost podman[112332]: 2025-11-27 09:11:39.415746831 +0000 UTC m=+0.079626372 container died c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vendor=Red Hat, Inc., config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, io.openshift.expose-services=, name=rhosp17/openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, tcib_managed=true, vcs-type=git, maintainer=OpenStack TripleO Team, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, batch=17.1_20251118.1, architecture=x86_64, distribution-scope=public, managed_by=tripleo_ansible, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, build-date=2025-11-19T00:35:22Z, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=nova_virtstoraged, version=17.1.12, com.redhat.component=openstack-nova-libvirt-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:11:39 localhost systemd[1]: tmp-crun.1u7cxd.mount: Deactivated successfully. Nov 27 04:11:39 localhost podman[112332]: 2025-11-27 09:11:39.464598435 +0000 UTC m=+0.128477956 container cleanup c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, tcib_managed=true, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, name=rhosp17/openstack-nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, build-date=2025-11-19T00:35:22Z, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, batch=17.1_20251118.1, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., release=1761123044, container_name=nova_virtstoraged, config_id=tripleo_step3, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, vcs-type=git, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, managed_by=tripleo_ansible, architecture=x86_64) Nov 27 04:11:39 localhost podman[112332]: nova_virtstoraged Nov 27 04:11:39 localhost podman[112346]: 2025-11-27 09:11:39.484722266 +0000 UTC m=+0.056617323 container cleanup c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, build-date=2025-11-19T00:35:22Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 nova-libvirt, io.buildah.version=1.41.4, name=rhosp17/openstack-nova-libvirt, io.openshift.expose-services=, com.redhat.component=openstack-nova-libvirt-container, release=1761123044, url=https://www.redhat.com, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, architecture=x86_64, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step3, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, distribution-scope=public, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, container_name=nova_virtstoraged, maintainer=OpenStack TripleO Team, version=17.1.12) Nov 27 04:11:39 localhost systemd[1]: libpod-conmon-c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4.scope: Deactivated successfully. Nov 27 04:11:39 localhost podman[112377]: error opening file `/run/crun/c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4/status`: No such file or directory Nov 27 04:11:39 localhost podman[112364]: 2025-11-27 09:11:39.569815763 +0000 UTC m=+0.058605767 container cleanup c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtstoraged, container_name=nova_virtstoraged, description=Red Hat OpenStack Platform 17.1 nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, vendor=Red Hat, Inc., name=rhosp17/openstack-nova-libvirt, tcib_managed=true, io.openshift.expose-services=, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, config_id=tripleo_step3, maintainer=OpenStack TripleO Team, com.redhat.component=openstack-nova-libvirt-container, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, release=1761123044, distribution-scope=public, managed_by=tripleo_ansible, vcs-type=git, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, batch=17.1_20251118.1, build-date=2025-11-19T00:35:22Z, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.41.4, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'cgroupns': 'host', 'depends_on': ['tripleo_nova_virtlogd_wrapper.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '4018231ef757a0bfd34181dc6dc2933a'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1', 'net': 'host', 'pid': 'host', 'pids_limit': 65536, 'privileged': True, 'restart': 'always', 'security_opt': ['label=level:s0', 'label=type:spc_t', 'label=filetype:container_file_t'], 'start_order': 3, 'ulimit': ['nofile=131072', 'nproc=126960'], 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/libvirt:/var/log/libvirt:shared,z', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/run:/run', '/sys/fs/cgroup:/sys/fs/cgroup', '/sys/fs/selinux:/sys/fs/selinux', '/etc/selinux/config:/etc/selinux/config:ro', '/etc/libvirt:/etc/libvirt:shared', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/libvirt:/var/lib/libvirt:shared', '/var/cache/libvirt:/var/cache/libvirt:shared', '/var/lib/vhost_sockets:/var/lib/vhost_sockets', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/var/lib/kolla/config_files/nova_virtstoraged.json:/var/lib/kolla/config_files/config.json:ro']}) Nov 27 04:11:39 localhost podman[112364]: nova_virtstoraged Nov 27 04:11:39 localhost systemd[1]: tripleo_nova_virtstoraged.service: Deactivated successfully. Nov 27 04:11:39 localhost systemd[1]: Stopped nova_virtstoraged container. Nov 27 04:11:39 localhost systemd[1]: var-lib-containers-storage-overlay-ea384ad2fb84d831a2a161d3935eaef9cd88cfe43ab225d67f714d0d77b43ecf-merged.mount: Deactivated successfully. Nov 27 04:11:39 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c059ad281608c8b48be82447157af04fa9f6e9412b710a2ee5e7c4abb4675aa4-userdata-shm.mount: Deactivated successfully. Nov 27 04:11:40 localhost python3.9[112470]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_ovn_controller.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:11:40 localhost systemd[1]: Reloading. Nov 27 04:11:40 localhost systemd-rc-local-generator[112497]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:11:40 localhost systemd-sysv-generator[112502]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:11:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:11:40 localhost systemd[1]: Stopping ovn_controller container... Nov 27 04:11:40 localhost systemd[1]: libpod-aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.scope: Deactivated successfully. Nov 27 04:11:40 localhost systemd[1]: libpod-aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.scope: Consumed 2.827s CPU time. Nov 27 04:11:40 localhost podman[112511]: 2025-11-27 09:11:40.841159486 +0000 UTC m=+0.076337144 container died aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, name=rhosp17/openstack-ovn-controller, container_name=ovn_controller, version=17.1.12, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1761123044, tcib_managed=true, url=https://www.redhat.com, konflux.additional-tags=17.1.12 17.1_20251118.1, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, build-date=2025-11-18T23:34:05Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, vendor=Red Hat, Inc., com.redhat.component=openstack-ovn-controller-container, architecture=x86_64, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-type=git, io.openshift.expose-services=, batch=17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.buildah.version=1.41.4, distribution-scope=public, managed_by=tripleo_ansible) Nov 27 04:11:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.timer: Deactivated successfully. Nov 27 04:11:40 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2. Nov 27 04:11:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed to open /run/systemd/transient/aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: No such file or directory Nov 27 04:11:40 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2-userdata-shm.mount: Deactivated successfully. Nov 27 04:11:40 localhost podman[112511]: 2025-11-27 09:11:40.890902843 +0000 UTC m=+0.126080471 container cleanup aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, architecture=x86_64, summary=Red Hat OpenStack Platform 17.1 ovn-controller, batch=17.1_20251118.1, maintainer=OpenStack TripleO Team, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, distribution-scope=public, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, version=17.1.12, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, com.redhat.component=openstack-ovn-controller-container, io.buildah.version=1.41.4, build-date=2025-11-18T23:34:05Z, tcib_managed=true, url=https://www.redhat.com, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, vendor=Red Hat, Inc., config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, io.openshift.expose-services=, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, managed_by=tripleo_ansible, config_id=tripleo_step4) Nov 27 04:11:40 localhost podman[112511]: ovn_controller Nov 27 04:11:40 localhost systemd[1]: var-lib-containers-storage-overlay-6cf7f72aaf45e45d932f6befc623c64b0e80d1af7cd41be6a403137845b9774f-merged.mount: Deactivated successfully. Nov 27 04:11:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.timer: Failed to open /run/systemd/transient/aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.timer: No such file or directory Nov 27 04:11:40 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed to open /run/systemd/transient/aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: No such file or directory Nov 27 04:11:40 localhost podman[112525]: 2025-11-27 09:11:40.926484259 +0000 UTC m=+0.075744427 container cleanup aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, config_id=tripleo_step4, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.expose-services=, maintainer=OpenStack TripleO Team, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, summary=Red Hat OpenStack Platform 17.1 ovn-controller, build-date=2025-11-18T23:34:05Z, vcs-type=git, com.redhat.component=openstack-ovn-controller-container, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, managed_by=tripleo_ansible, version=17.1.12, tcib_managed=true, url=https://www.redhat.com, release=1761123044, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, konflux.additional-tags=17.1.12 17.1_20251118.1, name=rhosp17/openstack-ovn-controller, batch=17.1_20251118.1, architecture=x86_64, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=Red Hat OpenStack Platform 17.1 ovn-controller, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272) Nov 27 04:11:40 localhost systemd[1]: libpod-conmon-aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.scope: Deactivated successfully. Nov 27 04:11:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:11:41 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.timer: Failed to open /run/systemd/transient/aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.timer: No such file or directory Nov 27 04:11:41 localhost systemd[1]: aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: Failed to open /run/systemd/transient/aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2.service: No such file or directory Nov 27 04:11:41 localhost podman[112539]: 2025-11-27 09:11:41.044650807 +0000 UTC m=+0.085018597 container cleanup aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, distribution-scope=public, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, vcs-type=git, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, konflux.additional-tags=17.1.12 17.1_20251118.1, architecture=x86_64, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, release=1761123044, io.openshift.expose-services=, managed_by=tripleo_ansible, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_controller, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, tcib_managed=true, build-date=2025-11-18T23:34:05Z, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, name=rhosp17/openstack-ovn-controller, url=https://www.redhat.com, io.buildah.version=1.41.4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, vendor=Red Hat, Inc., cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, description=Red Hat OpenStack Platform 17.1 ovn-controller, maintainer=OpenStack TripleO Team, batch=17.1_20251118.1) Nov 27 04:11:41 localhost podman[112539]: ovn_controller Nov 27 04:11:41 localhost systemd[1]: tripleo_ovn_controller.service: Deactivated successfully. Nov 27 04:11:41 localhost systemd[1]: Stopped ovn_controller container. Nov 27 04:11:41 localhost systemd[1]: tmp-crun.X3XwKz.mount: Deactivated successfully. Nov 27 04:11:41 localhost podman[112550]: 2025-11-27 09:11:41.143177576 +0000 UTC m=+0.099162898 container health_status 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, health_status=unhealthy, konflux.additional-tags=17.1.12 17.1_20251118.1, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, architecture=x86_64, managed_by=tripleo_ansible, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, maintainer=OpenStack TripleO Team, container_name=ovn_metadata_agent, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn, release=1761123044, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, distribution-scope=public, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, url=https://www.redhat.com, vendor=Red Hat, Inc., io.buildah.version=1.41.4, io.openshift.expose-services=, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=tripleo_step4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, build-date=2025-11-19T00:14:25Z, tcib_managed=true) Nov 27 04:11:41 localhost podman[112550]: 2025-11-27 09:11:41.16082264 +0000 UTC m=+0.116807942 container exec_died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, config_id=tripleo_step4, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, tcib_managed=true, url=https://www.redhat.com, name=rhosp17/openstack-neutron-metadata-agent-ovn, vcs-type=git, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, batch=17.1_20251118.1, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, managed_by=tripleo_ansible, build-date=2025-11-19T00:14:25Z, container_name=ovn_metadata_agent, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, release=1761123044, io.buildah.version=1.41.4, architecture=x86_64, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, maintainer=OpenStack TripleO Team, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:11:41 localhost podman[112550]: unhealthy Nov 27 04:11:41 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:11:41 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed with result 'exit-code'. Nov 27 04:11:41 localhost python3.9[112661]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_ovn_metadata_agent.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:11:41 localhost systemd[1]: Reloading. Nov 27 04:11:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36078 DF PROTO=TCP SPT=56752 DPT=9882 SEQ=739974169 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F88DA0000000001030307) Nov 27 04:11:41 localhost systemd-rc-local-generator[112690]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:11:41 localhost systemd-sysv-generator[112694]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:11:41 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:11:42 localhost systemd[1]: Stopping ovn_metadata_agent container... Nov 27 04:11:42 localhost systemd[1]: libpod-2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.scope: Deactivated successfully. Nov 27 04:11:42 localhost systemd[1]: libpod-2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.scope: Consumed 11.413s CPU time. Nov 27 04:11:42 localhost podman[112702]: 2025-11-27 09:11:42.791709319 +0000 UTC m=+0.645602098 container died 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, batch=17.1_20251118.1, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, architecture=x86_64, distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, tcib_managed=true, io.openshift.expose-services=, release=1761123044, vcs-type=git, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, config_id=tripleo_step4, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, version=17.1.12, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn) Nov 27 04:11:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.timer: Deactivated successfully. Nov 27 04:11:42 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954. Nov 27 04:11:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed to open /run/systemd/transient/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: No such file or directory Nov 27 04:11:42 localhost systemd[1]: tmp-crun.dnu827.mount: Deactivated successfully. Nov 27 04:11:42 localhost podman[112702]: 2025-11-27 09:11:42.862822272 +0000 UTC m=+0.716715001 container cleanup 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, url=https://www.redhat.com, architecture=x86_64, maintainer=OpenStack TripleO Team, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, managed_by=tripleo_ansible, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-type=git, config_id=tripleo_step4, version=17.1.12, io.buildah.version=1.41.4, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, container_name=ovn_metadata_agent, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, tcib_managed=true, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, build-date=2025-11-19T00:14:25Z, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, release=1761123044, io.openshift.expose-services=, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, name=rhosp17/openstack-neutron-metadata-agent-ovn) Nov 27 04:11:42 localhost podman[112702]: ovn_metadata_agent Nov 27 04:11:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.timer: Failed to open /run/systemd/transient/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.timer: No such file or directory Nov 27 04:11:42 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed to open /run/systemd/transient/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: No such file or directory Nov 27 04:11:42 localhost podman[112715]: 2025-11-27 09:11:42.888338158 +0000 UTC m=+0.090405003 container cleanup 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, maintainer=OpenStack TripleO Team, config_id=tripleo_step4, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, release=1761123044, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, build-date=2025-11-19T00:14:25Z, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, url=https://www.redhat.com, version=17.1.12, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, name=rhosp17/openstack-neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, container_name=ovn_metadata_agent, batch=17.1_20251118.1, vcs-type=git, vendor=Red Hat, Inc., tcib_managed=true, distribution-scope=public) Nov 27 04:11:43 localhost systemd[1]: var-lib-containers-storage-overlay-f96fef7b313009df8cfca80855d5a7b323e2e769d6b440ee9ef0691972acd2bc-merged.mount: Deactivated successfully. Nov 27 04:11:43 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954-userdata-shm.mount: Deactivated successfully. Nov 27 04:11:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31133 DF PROTO=TCP SPT=35680 DPT=9102 SEQ=2250256347 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F97690000000001030307) Nov 27 04:11:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15743 DF PROTO=TCP SPT=55930 DPT=9100 SEQ=2797184436 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127F9F9A0000000001030307) Nov 27 04:11:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15744 DF PROTO=TCP SPT=55930 DPT=9100 SEQ=2797184436 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FAF5B0000000001030307) Nov 27 04:11:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54717 DF PROTO=TCP SPT=37556 DPT=9101 SEQ=748199281 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FB75A0000000001030307) Nov 27 04:11:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29543 DF PROTO=TCP SPT=40420 DPT=9105 SEQ=1683781505 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FC35A0000000001030307) Nov 27 04:11:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54719 DF PROTO=TCP SPT=37556 DPT=9101 SEQ=748199281 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FCF1A0000000001030307) Nov 27 04:12:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29545 DF PROTO=TCP SPT=40420 DPT=9105 SEQ=1683781505 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FDB1A0000000001030307) Nov 27 04:12:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54720 DF PROTO=TCP SPT=37556 DPT=9101 SEQ=748199281 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FEFDA0000000001030307) Nov 27 04:12:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4178 DF PROTO=TCP SPT=55780 DPT=9882 SEQ=1376659644 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FF2070000000001030307) Nov 27 04:12:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4180 DF PROTO=TCP SPT=55780 DPT=9882 SEQ=1376659644 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A127FFE1A0000000001030307) Nov 27 04:12:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64657 DF PROTO=TCP SPT=40898 DPT=9102 SEQ=4209071291 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12800C990000000001030307) Nov 27 04:12:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=3185 DF PROTO=TCP SPT=46224 DPT=9100 SEQ=3306723941 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128014DA0000000001030307) Nov 27 04:12:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=3186 DF PROTO=TCP SPT=46224 DPT=9100 SEQ=3306723941 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280249A0000000001030307) Nov 27 04:12:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21013 DF PROTO=TCP SPT=51102 DPT=9101 SEQ=3639706221 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12802C5A0000000001030307) Nov 27 04:12:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=65078 DF PROTO=TCP SPT=58638 DPT=9101 SEQ=2504457277 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128037DA0000000001030307) Nov 27 04:12:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31335 DF PROTO=TCP SPT=52230 DPT=9105 SEQ=969979810 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128043DA0000000001030307) Nov 27 04:12:30 localhost systemd[1]: tmp-crun.KJiLk1.mount: Deactivated successfully. Nov 27 04:12:30 localhost podman[112835]: 2025-11-27 09:12:30.915768532 +0000 UTC m=+0.102587570 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., io.buildah.version=1.33.12, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, vcs-type=git, release=553, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, ceph=True) Nov 27 04:12:31 localhost podman[112835]: 2025-11-27 09:12:31.02208479 +0000 UTC m=+0.208903868 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, version=7, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, build-date=2025-09-24T08:57:55, release=553, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , name=rhceph) Nov 27 04:12:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32302 DF PROTO=TCP SPT=56862 DPT=9105 SEQ=2670045307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280505A0000000001030307) Nov 27 04:12:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21016 DF PROTO=TCP SPT=51102 DPT=9101 SEQ=3639706221 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128063DA0000000001030307) Nov 27 04:12:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20831 DF PROTO=TCP SPT=50560 DPT=9882 SEQ=2658528013 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128067370000000001030307) Nov 27 04:12:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20833 DF PROTO=TCP SPT=50560 DPT=9882 SEQ=2658528013 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280735A0000000001030307) Nov 27 04:12:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=34989 DF PROTO=TCP SPT=57744 DPT=9102 SEQ=4004314090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128081C90000000001030307) Nov 27 04:12:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17743 DF PROTO=TCP SPT=47350 DPT=9100 SEQ=85884175 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128089DA0000000001030307) Nov 27 04:12:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17744 DF PROTO=TCP SPT=47350 DPT=9100 SEQ=85884175 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280999A0000000001030307) Nov 27 04:12:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40723 DF PROTO=TCP SPT=60856 DPT=9101 SEQ=3082387037 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280A19A0000000001030307) Nov 27 04:12:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=576 DF PROTO=TCP SPT=50450 DPT=9105 SEQ=3143163930 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280ADDB0000000001030307) Nov 27 04:12:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40725 DF PROTO=TCP SPT=60856 DPT=9101 SEQ=3082387037 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280B95B0000000001030307) Nov 27 04:13:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=578 DF PROTO=TCP SPT=50450 DPT=9105 SEQ=3143163930 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280C59A0000000001030307) Nov 27 04:13:06 localhost systemd[1]: tripleo_ovn_metadata_agent.service: State 'stop-sigterm' timed out. Killing. Nov 27 04:13:06 localhost systemd[1]: tripleo_ovn_metadata_agent.service: Killing process 73047 (conmon) with signal SIGKILL. Nov 27 04:13:06 localhost systemd[1]: tripleo_ovn_metadata_agent.service: Main process exited, code=killed, status=9/KILL Nov 27 04:13:06 localhost systemd[1]: libpod-conmon-2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.scope: Deactivated successfully. Nov 27 04:13:06 localhost systemd[1]: tmp-crun.zYjEbc.mount: Deactivated successfully. Nov 27 04:13:06 localhost podman[112988]: error opening file `/run/crun/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954/status`: No such file or directory Nov 27 04:13:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.timer: Failed to open /run/systemd/transient/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.timer: No such file or directory Nov 27 04:13:06 localhost systemd[1]: 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: Failed to open /run/systemd/transient/2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954.service: No such file or directory Nov 27 04:13:07 localhost podman[112977]: 2025-11-27 09:13:06.999568389 +0000 UTC m=+0.091156242 container cleanup 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, konflux.additional-tags=17.1.12 17.1_20251118.1, config_id=tripleo_step4, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, tcib_managed=true, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, io.openshift.expose-services=, url=https://www.redhat.com, managed_by=tripleo_ansible, name=rhosp17/openstack-neutron-metadata-agent-ovn, vendor=Red Hat, Inc., vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, vcs-type=git, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, distribution-scope=public, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, architecture=x86_64, release=1761123044, build-date=2025-11-19T00:14:25Z, maintainer=OpenStack TripleO Team, version=17.1.12, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, batch=17.1_20251118.1) Nov 27 04:13:07 localhost podman[112977]: ovn_metadata_agent Nov 27 04:13:07 localhost systemd[1]: tripleo_ovn_metadata_agent.service: Failed with result 'timeout'. Nov 27 04:13:07 localhost systemd[1]: Stopped ovn_metadata_agent container. Nov 27 04:13:07 localhost python3.9[113082]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_rsyslog.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:13:07 localhost systemd[1]: Reloading. Nov 27 04:13:07 localhost systemd-rc-local-generator[113111]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:13:07 localhost systemd-sysv-generator[113115]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:13:07 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:13:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40726 DF PROTO=TCP SPT=60856 DPT=9101 SEQ=3082387037 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280D9DA0000000001030307) Nov 27 04:13:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23501 DF PROTO=TCP SPT=60522 DPT=9882 SEQ=3479543601 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280DC670000000001030307) Nov 27 04:13:09 localhost python3.9[113212]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_ceilometer_agent_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:10 localhost python3.9[113304]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_ceilometer_agent_ipmi.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:10 localhost python3.9[113396]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_collectd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:11 localhost python3.9[113488]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_iscsid.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23503 DF PROTO=TCP SPT=60522 DPT=9882 SEQ=3479543601 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280E85A0000000001030307) Nov 27 04:13:11 localhost python3.9[113580]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_logrotate_crond.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:12 localhost python3.9[113672]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_metrics_qdr.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:12 localhost python3.9[113764]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_neutron_dhcp.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:13 localhost python3.9[113856]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_neutron_l3_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:14 localhost python3.9[113948]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_neutron_ovs_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:14 localhost python3.9[114040]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:15 localhost python3.9[114132]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21562 DF PROTO=TCP SPT=35438 DPT=9102 SEQ=1300606912 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280F6F90000000001030307) Nov 27 04:13:15 localhost python3.9[114224]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtlogd_wrapper.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:16 localhost python3.9[114316]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtnodedevd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:17 localhost python3.9[114408]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtproxyd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59516 DF PROTO=TCP SPT=55448 DPT=9100 SEQ=2249105590 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1280FF1A0000000001030307) Nov 27 04:13:17 localhost python3.9[114500]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtqemud.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:18 localhost python3.9[114592]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtqemud_recover.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:18 localhost python3.9[114684]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtsecretd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:19 localhost python3.9[114776]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtstoraged.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:20 localhost python3.9[114868]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_ovn_controller.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:20 localhost python3.9[114960]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_ovn_metadata_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:21 localhost python3.9[115052]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_rsyslog.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59517 DF PROTO=TCP SPT=55448 DPT=9100 SEQ=2249105590 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12810EDA0000000001030307) Nov 27 04:13:22 localhost python3.9[115144]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:23 localhost python3.9[115236]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_ipmi.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53814 DF PROTO=TCP SPT=46612 DPT=9101 SEQ=665799772 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128116DA0000000001030307) Nov 27 04:13:23 localhost python3.9[115328]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_collectd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:24 localhost python3.9[115420]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_iscsid.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:24 localhost python3.9[115512]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_logrotate_crond.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:25 localhost python3.9[115604]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_metrics_qdr.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:26 localhost python3.9[115696]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_neutron_dhcp.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:26 localhost python3.9[115788]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_neutron_l3_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14387 DF PROTO=TCP SPT=44920 DPT=9105 SEQ=4109056765 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128122DB0000000001030307) Nov 27 04:13:27 localhost python3.9[115880]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_neutron_ovs_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:27 localhost python3.9[115972]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:28 localhost python3.9[116064]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:28 localhost python3.9[116156]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtlogd_wrapper.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:29 localhost python3.9[116248]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtnodedevd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32305 DF PROTO=TCP SPT=56862 DPT=9105 SEQ=2670045307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12812DDA0000000001030307) Nov 27 04:13:29 localhost python3.9[116340]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtproxyd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:30 localhost python3.9[116432]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:31 localhost python3.9[116524]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud_recover.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:31 localhost python3.9[116616]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtsecretd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:32 localhost python3.9[116708]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtstoraged.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:32 localhost python3.9[116800]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_ovn_controller.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14389 DF PROTO=TCP SPT=44920 DPT=9105 SEQ=4109056765 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12813A9A0000000001030307) Nov 27 04:13:33 localhost python3.9[116922]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_ovn_metadata_agent.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:33 localhost python3.9[117047]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_rsyslog.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:13:35 localhost python3.9[117154]: ansible-ansible.legacy.command Invoked with _raw_params=if systemctl is-active certmonger.service; then#012 systemctl disable --now certmonger.service#012 test -f /etc/systemd/system/certmonger.service || systemctl mask certmonger.service#012fi#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:36 localhost python3.9[117246]: ansible-ansible.builtin.find Invoked with file_type=any hidden=True paths=['/var/lib/certmonger/requests'] patterns=[] read_whole_file=False age_stamp=mtime recurse=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:13:37 localhost python3.9[117338]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:13:37 localhost systemd[1]: Reloading. Nov 27 04:13:37 localhost systemd-rc-local-generator[117361]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:13:37 localhost systemd-sysv-generator[117364]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:13:37 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:13:38 localhost python3.9[117466]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_ceilometer_agent_compute.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53817 DF PROTO=TCP SPT=46612 DPT=9101 SEQ=665799772 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12814FDA0000000001030307) Nov 27 04:13:38 localhost python3.9[117559]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_ceilometer_agent_ipmi.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=30260 DF PROTO=TCP SPT=50236 DPT=9882 SEQ=3835165229 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128151960000000001030307) Nov 27 04:13:39 localhost python3.9[117652]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_collectd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:39 localhost python3.9[117745]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_iscsid.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:41 localhost python3.9[117838]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_logrotate_crond.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=30262 DF PROTO=TCP SPT=50236 DPT=9882 SEQ=3835165229 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12815D9A0000000001030307) Nov 27 04:13:43 localhost python3.9[117931]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_metrics_qdr.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:43 localhost python3.9[118024]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_neutron_dhcp.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:44 localhost python3.9[118117]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_neutron_l3_agent.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:44 localhost python3.9[118210]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_neutron_ovs_agent.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:45 localhost python3.9[118303]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_compute.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13997 DF PROTO=TCP SPT=46766 DPT=9102 SEQ=216925109 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12816C290000000001030307) Nov 27 04:13:46 localhost python3.9[118396]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_migration_target.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:46 localhost python3.9[118489]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtlogd_wrapper.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:47 localhost python3.9[118582]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtnodedevd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31027 DF PROTO=TCP SPT=33370 DPT=9100 SEQ=1176840498 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281745B0000000001030307) Nov 27 04:13:47 localhost python3.9[118675]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtproxyd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:48 localhost python3.9[118768]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtqemud.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:49 localhost python3.9[118861]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtqemud_recover.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:49 localhost python3.9[118954]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtsecretd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:50 localhost python3.9[119047]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtstoraged.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:50 localhost python3.9[119140]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_ovn_controller.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31028 DF PROTO=TCP SPT=33370 DPT=9100 SEQ=1176840498 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281841B0000000001030307) Nov 27 04:13:52 localhost python3.9[119233]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_ovn_metadata_agent.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:53 localhost python3.9[119326]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_rsyslog.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:13:53 localhost systemd[1]: session-38.scope: Deactivated successfully. Nov 27 04:13:53 localhost systemd[1]: session-38.scope: Consumed 48.880s CPU time. Nov 27 04:13:53 localhost systemd-logind[761]: Session 38 logged out. Waiting for processes to exit. Nov 27 04:13:53 localhost systemd-logind[761]: Removed session 38. Nov 27 04:13:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43964 DF PROTO=TCP SPT=60998 DPT=9101 SEQ=2834469497 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12818C1A0000000001030307) Nov 27 04:13:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40728 DF PROTO=TCP SPT=60856 DPT=9101 SEQ=3082387037 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128197DA0000000001030307) Nov 27 04:13:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=581 DF PROTO=TCP SPT=50450 DPT=9105 SEQ=3143163930 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281A3DA0000000001030307) Nov 27 04:14:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37022 DF PROTO=TCP SPT=51274 DPT=9105 SEQ=2460320816 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281AFDA0000000001030307) Nov 27 04:14:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43967 DF PROTO=TCP SPT=60998 DPT=9101 SEQ=2834469497 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281C3DA0000000001030307) Nov 27 04:14:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14765 DF PROTO=TCP SPT=37084 DPT=9882 SEQ=3598920708 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281C6C70000000001030307) Nov 27 04:14:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14767 DF PROTO=TCP SPT=37084 DPT=9882 SEQ=3598920708 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281D2DA0000000001030307) Nov 27 04:14:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11688 DF PROTO=TCP SPT=51368 DPT=9102 SEQ=1460988909 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281E1590000000001030307) Nov 27 04:14:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31978 DF PROTO=TCP SPT=37658 DPT=9100 SEQ=426786368 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281E99B0000000001030307) Nov 27 04:14:17 localhost sshd[119343]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:14:18 localhost systemd-logind[761]: New session 39 of user zuul. Nov 27 04:14:18 localhost systemd[1]: Started Session 39 of User zuul. Nov 27 04:14:18 localhost python3.9[119436]: ansible-ansible.legacy.ping Invoked with data=pong Nov 27 04:14:20 localhost python3.9[119540]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:14:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31979 DF PROTO=TCP SPT=37658 DPT=9100 SEQ=426786368 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1281F95A0000000001030307) Nov 27 04:14:21 localhost python3.9[119632]: ansible-ansible.legacy.command Invoked with _raw_params=PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin which growvols#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:14:22 localhost python3.9[119725]: ansible-ansible.builtin.stat Invoked with path=/etc/ansible/facts.d/bootc.fact follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:14:23 localhost python3.9[119817]: ansible-ansible.builtin.file Invoked with mode=755 path=/etc/ansible/facts.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:14:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38739 DF PROTO=TCP SPT=39660 DPT=9101 SEQ=3533785877 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282011B0000000001030307) Nov 27 04:14:24 localhost python3.9[119909]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/bootc.fact follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:14:24 localhost python3.9[119982]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/bootc.fact mode=755 src=/home/zuul/.ansible/tmp/ansible-tmp-1764234863.78586-180-239105097392573/.source.fact _original_basename=bootc.fact follow=False checksum=eb4122ce7fc50a38407beb511c4ff8c178005b12 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:14:25 localhost python3.9[120074]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:14:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6435 DF PROTO=TCP SPT=56254 DPT=9105 SEQ=2391378953 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12820D5A0000000001030307) Nov 27 04:14:26 localhost python3.9[120170]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/var/log/journal setype=var_log_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:14:27 localhost python3.9[120262]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/config-data/ansible-generated recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:14:28 localhost python3.9[120352]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:14:29 localhost network[120369]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:14:29 localhost network[120370]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:14:29 localhost network[120371]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:14:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38741 DF PROTO=TCP SPT=39660 DPT=9101 SEQ=3533785877 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128218DA0000000001030307) Nov 27 04:14:30 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:14:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6437 DF PROTO=TCP SPT=56254 DPT=9105 SEQ=2391378953 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282251B0000000001030307) Nov 27 04:14:32 localhost python3.9[120568]: ansible-ansible.builtin.lineinfile Invoked with line=cloud-init=disabled path=/proc/cmdline state=present encoding=utf-8 backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:14:33 localhost python3.9[120658]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:14:34 localhost python3.9[120784]: ansible-ansible.legacy.command Invoked with _raw_params=# This is a hack to deploy RDO Delorean repos to RHEL as if it were Centos 9 Stream#012set -euxo pipefail#012curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz#012python3 -m venv ./venv#012PBR_VERSION=0.0.0 ./venv/bin/pip install ./repo-setup-main#012# This is required for FIPS enabled until trunk.rdoproject.org#012# is not being served from a centos7 host, tracked by#012# https://issues.redhat.com/browse/RHOSZUUL-1517#012dnf -y install crypto-policies#012update-crypto-policies --set FIPS:NO-ENFORCE-EMS#012./venv/bin/repo-setup current-podified -b antelope -d centos9 --stream#012#012# Exclude ceph-common-18.2.7 as it's pulling newer openssl not compatible#012# with rhel 9.2 openssh#012dnf config-manager --setopt centos9-storage.exclude="ceph-common-18.2.7" --save#012# FIXME: perform dnf upgrade for other packages in EDPM ansible#012# here we only ensuring that decontainerized libvirt can start#012dnf -y upgrade openstack-selinux#012rm -f /run/virtlogd.pid#012#012rm -rf repo-setup-main#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:14:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38742 DF PROTO=TCP SPT=39660 DPT=9101 SEQ=3533785877 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128239DA0000000001030307) Nov 27 04:14:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51722 DF PROTO=TCP SPT=45330 DPT=9882 SEQ=3682975349 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12823BF70000000001030307) Nov 27 04:14:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51724 DF PROTO=TCP SPT=45330 DPT=9882 SEQ=3682975349 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282481A0000000001030307) Nov 27 04:14:44 localhost systemd[1]: Stopping OpenSSH server daemon... Nov 27 04:14:44 localhost systemd[1]: sshd.service: Deactivated successfully. Nov 27 04:14:44 localhost systemd[1]: Stopped OpenSSH server daemon. Nov 27 04:14:44 localhost systemd[1]: Stopped target sshd-keygen.target. Nov 27 04:14:44 localhost systemd[1]: Stopping sshd-keygen.target... Nov 27 04:14:44 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:44 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:44 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:44 localhost systemd[1]: Reached target sshd-keygen.target. Nov 27 04:14:44 localhost systemd[1]: Starting OpenSSH server daemon... Nov 27 04:14:44 localhost sshd[120873]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:14:44 localhost systemd[1]: Started OpenSSH server daemon. Nov 27 04:14:44 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 04:14:44 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 04:14:44 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 04:14:44 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 04:14:44 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 04:14:44 localhost systemd[1]: run-ra61228a034834f27aacecd422277c167.service: Deactivated successfully. Nov 27 04:14:44 localhost systemd[1]: run-rba472c8bd3a94489b4973dec92ff1e42.service: Deactivated successfully. Nov 27 04:14:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37975 DF PROTO=TCP SPT=33900 DPT=9102 SEQ=2229541026 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282568B0000000001030307) Nov 27 04:14:45 localhost systemd[1]: Stopping OpenSSH server daemon... Nov 27 04:14:45 localhost systemd[1]: sshd.service: Deactivated successfully. Nov 27 04:14:45 localhost systemd[1]: Stopped OpenSSH server daemon. Nov 27 04:14:45 localhost systemd[1]: Stopped target sshd-keygen.target. Nov 27 04:14:45 localhost systemd[1]: Stopping sshd-keygen.target... Nov 27 04:14:45 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:45 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:45 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:14:45 localhost systemd[1]: Reached target sshd-keygen.target. Nov 27 04:14:45 localhost systemd[1]: Starting OpenSSH server daemon... Nov 27 04:14:45 localhost sshd[121241]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:14:45 localhost systemd[1]: Started OpenSSH server daemon. Nov 27 04:14:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64802 DF PROTO=TCP SPT=43306 DPT=9100 SEQ=3871037915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12825E9A0000000001030307) Nov 27 04:14:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64803 DF PROTO=TCP SPT=43306 DPT=9100 SEQ=3871037915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12826E5A0000000001030307) Nov 27 04:14:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28773 DF PROTO=TCP SPT=49566 DPT=9101 SEQ=2974904086 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282765A0000000001030307) Nov 27 04:14:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43969 DF PROTO=TCP SPT=60998 DPT=9101 SEQ=2834469497 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128281DA0000000001030307) Nov 27 04:14:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64804 DF PROTO=TCP SPT=43306 DPT=9100 SEQ=3871037915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12828DDA0000000001030307) Nov 27 04:15:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13608 DF PROTO=TCP SPT=33260 DPT=9105 SEQ=2026480090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12829A5A0000000001030307) Nov 27 04:15:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28776 DF PROTO=TCP SPT=49566 DPT=9101 SEQ=2974904086 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282ADDB0000000001030307) Nov 27 04:15:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12498 DF PROTO=TCP SPT=44456 DPT=9882 SEQ=583734451 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282B1280000000001030307) Nov 27 04:15:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12500 DF PROTO=TCP SPT=44456 DPT=9882 SEQ=583734451 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282BD1A0000000001030307) Nov 27 04:15:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63407 DF PROTO=TCP SPT=52922 DPT=9102 SEQ=2005051833 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282CBB80000000001030307) Nov 27 04:15:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13751 DF PROTO=TCP SPT=58366 DPT=9100 SEQ=2598966012 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282D3DB0000000001030307) Nov 27 04:15:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13752 DF PROTO=TCP SPT=58366 DPT=9100 SEQ=2598966012 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282E39A0000000001030307) Nov 27 04:15:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27249 DF PROTO=TCP SPT=33638 DPT=9101 SEQ=114828090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282EB9A0000000001030307) Nov 27 04:15:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41689 DF PROTO=TCP SPT=53848 DPT=9105 SEQ=3288991308 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1282F79A0000000001030307) Nov 27 04:15:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27251 DF PROTO=TCP SPT=33638 DPT=9101 SEQ=114828090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283035A0000000001030307) Nov 27 04:15:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41691 DF PROTO=TCP SPT=53848 DPT=9105 SEQ=3288991308 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12830F5B0000000001030307) Nov 27 04:15:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27252 DF PROTO=TCP SPT=33638 DPT=9101 SEQ=114828090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128323DB0000000001030307) Nov 27 04:15:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17241 DF PROTO=TCP SPT=53558 DPT=9882 SEQ=3107821286 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128326570000000001030307) Nov 27 04:15:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=17243 DF PROTO=TCP SPT=53558 DPT=9882 SEQ=3107821286 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283325B0000000001030307) Nov 27 04:15:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1726 DF PROTO=TCP SPT=44372 DPT=9102 SEQ=1352406766 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128340EA0000000001030307) Nov 27 04:15:46 localhost sshd[121700]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:15:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46680 DF PROTO=TCP SPT=43146 DPT=9100 SEQ=2218310706 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283491A0000000001030307) Nov 27 04:15:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46681 DF PROTO=TCP SPT=43146 DPT=9100 SEQ=2218310706 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128358DA0000000001030307) Nov 27 04:15:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32287 DF PROTO=TCP SPT=47410 DPT=9101 SEQ=3518476121 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128360DB0000000001030307) Nov 27 04:15:55 localhost kernel: SELinux: Converting 2755 SID table entries... Nov 27 04:15:55 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:15:55 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:15:55 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:15:55 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:15:55 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:15:55 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:15:55 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:15:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47835 DF PROTO=TCP SPT=36650 DPT=9105 SEQ=1333805747 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12836CDD0000000001030307) Nov 27 04:15:59 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=17 res=1 Nov 27 04:15:59 localhost python3.9[121900]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/ansible/facts.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:15:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13611 DF PROTO=TCP SPT=33260 DPT=9105 SEQ=2026480090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128377DB0000000001030307) Nov 27 04:16:00 localhost python3.9[121992]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/edpm.fact follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:16:00 localhost python3.9[122065]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/edpm.fact mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764234959.7606938-429-276590190464556/.source.fact _original_basename=.z3y3wcp2 follow=False checksum=03aee63dcf9b49b0ac4473b2f1a1b5d3783aa639 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:16:01 localhost python3.9[122155]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local', 'distribution'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:16:02 localhost python3.9[122253]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:16:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47837 DF PROTO=TCP SPT=36650 DPT=9105 SEQ=1333805747 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283849A0000000001030307) Nov 27 04:16:03 localhost python3.9[122307]: ansible-ansible.legacy.dnf Invoked with name=['driverctl', 'lvm2', 'crudini', 'jq', 'nftables', 'NetworkManager', 'openstack-selinux', 'python3-libselinux', 'python3-pyyaml', 'rsync', 'tmpwatch', 'sysstat', 'iproute-tc', 'ksmtuned', 'systemd-container', 'crypto-policies-scripts', 'grubby', 'sos'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:16:07 localhost systemd[1]: Reloading. Nov 27 04:16:07 localhost systemd-sysv-generator[122346]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:16:07 localhost systemd-rc-local-generator[122339]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:16:07 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:16:07 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 04:16:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32290 DF PROTO=TCP SPT=47410 DPT=9101 SEQ=3518476121 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128399DA0000000001030307) Nov 27 04:16:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33909 DF PROTO=TCP SPT=34120 DPT=9882 SEQ=173842644 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12839B870000000001030307) Nov 27 04:16:09 localhost python3.9[122447]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -V driverctl lvm2 crudini jq nftables NetworkManager openstack-selinux python3-libselinux python3-pyyaml rsync tmpwatch sysstat iproute-tc ksmtuned systemd-container crypto-policies-scripts grubby sos _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:16:11 localhost python3.9[122686]: ansible-ansible.posix.selinux Invoked with policy=targeted state=enforcing configfile=/etc/selinux/config update_kernel_param=False Nov 27 04:16:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33911 DF PROTO=TCP SPT=34120 DPT=9882 SEQ=173842644 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283A79B0000000001030307) Nov 27 04:16:11 localhost python3.9[122778]: ansible-ansible.legacy.command Invoked with cmd=dd if=/dev/zero of=/swap count=1024 bs=1M creates=/swap _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None removes=None stdin=None Nov 27 04:16:13 localhost python3.9[122871]: ansible-ansible.builtin.file Invoked with group=root mode=0600 owner=root path=/swap recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False state=None _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:16:14 localhost python3.9[122963]: ansible-ansible.posix.mount Invoked with dump=0 fstype=swap name=none opts=sw passno=0 src=/swap state=present path=none boot=True opts_no_log=False backup=False fstab=None Nov 27 04:16:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47683 DF PROTO=TCP SPT=60666 DPT=9102 SEQ=3891212132 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283B6190000000001030307) Nov 27 04:16:15 localhost python3.9[123055]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/pki/ca-trust/source/anchors setype=cert_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:16:16 localhost python3.9[123147]: ansible-ansible.legacy.stat Invoked with path=/etc/pki/ca-trust/source/anchors/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:16:16 localhost python3.9[123220]: ansible-ansible.legacy.copy Invoked with dest=/etc/pki/ca-trust/source/anchors/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764234975.8091176-753-182345829590109/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:16:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8773 DF PROTO=TCP SPT=36950 DPT=9100 SEQ=3866503429 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283BE5B0000000001030307) Nov 27 04:16:18 localhost python3.9[123312]: ansible-ansible.builtin.stat Invoked with path=/etc/lvm/devices/system.devices follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:16:19 localhost python3.9[123406]: ansible-ansible.builtin.getent Invoked with database=passwd key=qemu fail_key=True service=None split=None Nov 27 04:16:21 localhost python3.9[123499]: ansible-ansible.builtin.getent Invoked with database=passwd key=hugetlbfs fail_key=True service=None split=None Nov 27 04:16:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8774 DF PROTO=TCP SPT=36950 DPT=9100 SEQ=3866503429 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283CE1B0000000001030307) Nov 27 04:16:22 localhost python3.9[123592]: ansible-ansible.builtin.group Invoked with gid=42477 name=hugetlbfs state=present force=False system=False local=False non_unique=False gid_min=None gid_max=None Nov 27 04:16:23 localhost python3.9[123690]: ansible-ansible.builtin.file Invoked with group=qemu mode=0755 owner=qemu path=/var/lib/vhost_sockets setype=virt_cache_t seuser=system_u state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None serole=None selevel=None attributes=None Nov 27 04:16:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31681 DF PROTO=TCP SPT=35840 DPT=9101 SEQ=970903645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283D5DB0000000001030307) Nov 27 04:16:24 localhost python3.9[123782]: ansible-ansible.legacy.dnf Invoked with name=['dracut-config-generic'] state=absent allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:16:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27254 DF PROTO=TCP SPT=33638 DPT=9101 SEQ=114828090 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283E1DA0000000001030307) Nov 27 04:16:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31683 DF PROTO=TCP SPT=35840 DPT=9101 SEQ=970903645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283ED9A0000000001030307) Nov 27 04:16:31 localhost python3.9[123876]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/modules-load.d setype=etc_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:16:32 localhost python3.9[123969]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/99-edpm.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:16:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37733 DF PROTO=TCP SPT=42568 DPT=9105 SEQ=1567917718 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1283F9DA0000000001030307) Nov 27 04:16:33 localhost python3.9[124042]: ansible-ansible.legacy.copy Invoked with dest=/etc/modules-load.d/99-edpm.conf group=root mode=0644 owner=root setype=etc_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764234991.8427773-1026-233350789964372/.source.conf follow=False _original_basename=edpm-modprobe.conf.j2 checksum=8021efe01721d8fa8cab46b95c00ec1be6dbb9d0 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:16:34 localhost python3.9[124134]: ansible-ansible.builtin.systemd Invoked with name=systemd-modules-load.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:16:34 localhost systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 27 04:16:34 localhost systemd[1]: Stopped Load Kernel Modules. Nov 27 04:16:34 localhost systemd[1]: Stopping Load Kernel Modules... Nov 27 04:16:34 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 04:16:34 localhost systemd-modules-load[124138]: Module 'msr' is built in Nov 27 04:16:34 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 04:16:35 localhost python3.9[124230]: ansible-ansible.legacy.stat Invoked with path=/etc/sysctl.d/99-edpm.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:16:35 localhost python3.9[124303]: ansible-ansible.legacy.copy Invoked with dest=/etc/sysctl.d/99-edpm.conf group=root mode=0644 owner=root setype=etc_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764234994.7968564-1095-156226329514517/.source.conf follow=False _original_basename=edpm-sysctl.conf.j2 checksum=2a366439721b855adcfe4d7f152babb68596a007 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:16:36 localhost python3.9[124395]: ansible-ansible.legacy.dnf Invoked with name=['tuned', 'tuned-profiles-cpu-partitioning'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:16:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31684 DF PROTO=TCP SPT=35840 DPT=9101 SEQ=970903645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12840DDA0000000001030307) Nov 27 04:16:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2258 DF PROTO=TCP SPT=43904 DPT=9882 SEQ=1027954867 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128410B70000000001030307) Nov 27 04:16:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2260 DF PROTO=TCP SPT=43904 DPT=9882 SEQ=1027954867 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12841CDB0000000001030307) Nov 27 04:16:43 localhost python3.9[124564]: ansible-ansible.builtin.stat Invoked with path=/etc/tuned/active_profile follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:16:44 localhost python3.9[124656]: ansible-ansible.builtin.slurp Invoked with src=/etc/tuned/active_profile Nov 27 04:16:45 localhost python3.9[124746]: ansible-ansible.builtin.stat Invoked with path=/etc/tuned/throughput-performance-variables.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:16:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15454 DF PROTO=TCP SPT=49638 DPT=9102 SEQ=2241950233 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12842B490000000001030307) Nov 27 04:16:46 localhost python3.9[124838]: ansible-ansible.builtin.systemd Invoked with enabled=True name=tuned state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:16:46 localhost systemd[1]: Stopping Dynamic System Tuning Daemon... Nov 27 04:16:46 localhost systemd[1]: tuned.service: Deactivated successfully. Nov 27 04:16:46 localhost systemd[1]: Stopped Dynamic System Tuning Daemon. Nov 27 04:16:46 localhost systemd[1]: tuned.service: Consumed 1.791s CPU time, no IO. Nov 27 04:16:46 localhost systemd[1]: Starting Dynamic System Tuning Daemon... Nov 27 04:16:47 localhost systemd[1]: Started Dynamic System Tuning Daemon. Nov 27 04:16:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40675 DF PROTO=TCP SPT=49380 DPT=9100 SEQ=1463594520 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284335A0000000001030307) Nov 27 04:16:48 localhost python3.9[124940]: ansible-ansible.builtin.slurp Invoked with src=/proc/cmdline Nov 27 04:16:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40676 DF PROTO=TCP SPT=49380 DPT=9100 SEQ=1463594520 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284431A0000000001030307) Nov 27 04:16:52 localhost python3.9[125032]: ansible-ansible.builtin.systemd Invoked with enabled=False name=ksm.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:16:53 localhost systemd[1]: Reloading. Nov 27 04:16:53 localhost systemd-sysv-generator[125062]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:16:53 localhost systemd-rc-local-generator[125058]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:16:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:16:53 localhost systemd[1]: Starting dnf makecache... Nov 27 04:16:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6889 DF PROTO=TCP SPT=36260 DPT=9101 SEQ=58049466 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12844B1A0000000001030307) Nov 27 04:16:53 localhost dnf[125071]: Updating Subscription Management repositories. Nov 27 04:16:54 localhost python3.9[125163]: ansible-ansible.builtin.systemd Invoked with enabled=False name=ksmtuned.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:16:54 localhost systemd[1]: Reloading. Nov 27 04:16:54 localhost systemd-sysv-generator[125196]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:16:54 localhost systemd-rc-local-generator[125192]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:16:54 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:16:55 localhost python3.9[125293]: ansible-ansible.legacy.command Invoked with _raw_params=mkswap "/swap" _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:16:55 localhost dnf[125071]: Metadata cache refreshed recently. Nov 27 04:16:55 localhost systemd[1]: dnf-makecache.service: Deactivated successfully. Nov 27 04:16:55 localhost systemd[1]: Finished dnf makecache. Nov 27 04:16:55 localhost systemd[1]: dnf-makecache.service: Consumed 2.047s CPU time. Nov 27 04:16:55 localhost python3.9[125386]: ansible-ansible.legacy.command Invoked with _raw_params=swapon "/swap" _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:16:56 localhost kernel: Adding 1048572k swap on /swap. Priority:-2 extents:1 across:1048572k FS Nov 27 04:16:56 localhost python3.9[125479]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/bin/update-ca-trust _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:16:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35361 DF PROTO=TCP SPT=44218 DPT=9105 SEQ=3959886353 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284575A0000000001030307) Nov 27 04:16:58 localhost python3.9[125578]: ansible-ansible.legacy.command Invoked with _raw_params=echo 2 >/sys/kernel/mm/ksm/run _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:16:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6891 DF PROTO=TCP SPT=36260 DPT=9101 SEQ=58049466 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128462DA0000000001030307) Nov 27 04:16:59 localhost python3.9[125671]: ansible-ansible.builtin.systemd Invoked with name=systemd-sysctl.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:17:00 localhost systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 27 04:17:00 localhost systemd[1]: Stopped Apply Kernel Variables. Nov 27 04:17:00 localhost systemd[1]: Stopping Apply Kernel Variables... Nov 27 04:17:00 localhost systemd[1]: Starting Apply Kernel Variables... Nov 27 04:17:00 localhost systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 27 04:17:00 localhost systemd[1]: Finished Apply Kernel Variables. Nov 27 04:17:01 localhost systemd[1]: session-39.scope: Deactivated successfully. Nov 27 04:17:01 localhost systemd[1]: session-39.scope: Consumed 1min 56.277s CPU time. Nov 27 04:17:01 localhost systemd-logind[761]: Session 39 logged out. Waiting for processes to exit. Nov 27 04:17:01 localhost systemd-logind[761]: Removed session 39. Nov 27 04:17:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35363 DF PROTO=TCP SPT=44218 DPT=9105 SEQ=3959886353 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12846F1B0000000001030307) Nov 27 04:17:07 localhost sshd[125691]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:17:07 localhost systemd-logind[761]: New session 40 of user zuul. Nov 27 04:17:07 localhost systemd[1]: Started Session 40 of User zuul. Nov 27 04:17:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6892 DF PROTO=TCP SPT=36260 DPT=9101 SEQ=58049466 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128483DA0000000001030307) Nov 27 04:17:08 localhost python3.9[125784]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:17:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11311 DF PROTO=TCP SPT=34104 DPT=9882 SEQ=1140350788 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128485E70000000001030307) Nov 27 04:17:09 localhost python3.9[125878]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:17:11 localhost python3.9[125974]: ansible-ansible.legacy.command Invoked with _raw_params=PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin which growvols#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:17:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11313 DF PROTO=TCP SPT=34104 DPT=9882 SEQ=1140350788 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128491DA0000000001030307) Nov 27 04:17:12 localhost python3.9[126065]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:17:13 localhost python3.9[126161]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:17:14 localhost python3.9[126215]: ansible-ansible.legacy.dnf Invoked with name=['podman'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:17:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40318 DF PROTO=TCP SPT=37630 DPT=9102 SEQ=503558255 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284A0790000000001030307) Nov 27 04:17:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14390 DF PROTO=TCP SPT=58830 DPT=9100 SEQ=2774045645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284A89A0000000001030307) Nov 27 04:17:18 localhost python3.9[126309]: ansible-ansible.builtin.setup Invoked with filter=['ansible_interfaces'] gather_subset=['!all', '!min', 'network'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:17:19 localhost python3.9[126464]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/containers/networks recurse=True state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:17:21 localhost python3.9[126556]: ansible-ansible.legacy.command Invoked with _raw_params=podman network inspect podman#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:17:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14391 DF PROTO=TCP SPT=58830 DPT=9100 SEQ=2774045645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284B85B0000000001030307) Nov 27 04:17:22 localhost python3.9[126660]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/networks/podman.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:17:22 localhost python3.9[126708]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/containers/networks/podman.json _original_basename=podman_network_config.j2 recurse=False state=file path=/etc/containers/networks/podman.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:17:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59392 DF PROTO=TCP SPT=39920 DPT=9101 SEQ=3991612988 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284C05A0000000001030307) Nov 27 04:17:23 localhost python3.9[126800]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/20-edpm-podman-registries.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:17:24 localhost python3.9[126873]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/registries.conf.d/20-edpm-podman-registries.conf group=root mode=0644 owner=root setype=etc_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235043.358801-326-167265888752017/.source.conf follow=False _original_basename=registries.conf.j2 checksum=804a0d01b832e60d20f779a331306df708c87b02 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:17:25 localhost python3.9[126965]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=pids_limit owner=root path=/etc/containers/containers.conf section=containers setype=etc_t value=4096 backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:17:25 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.4 (251 of 333 items), suggesting rotation. Nov 27 04:17:25 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:17:25 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:17:25 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:17:26 localhost python3.9[127058]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=events_logger owner=root path=/etc/containers/containers.conf section=engine setype=etc_t value="journald" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:17:26 localhost python3.9[127150]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=runtime owner=root path=/etc/containers/containers.conf section=engine setype=etc_t value="crun" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:17:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31686 DF PROTO=TCP SPT=35840 DPT=9101 SEQ=970903645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284CBDB0000000001030307) Nov 27 04:17:27 localhost python3.9[127242]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=network_backend owner=root path=/etc/containers/containers.conf section=network setype=etc_t value="netavark" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:17:28 localhost python3.9[127332]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'distribution'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:17:29 localhost python3.9[127426]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['driverctl', 'lvm2', 'crudini', 'jq', 'nftables', 'NetworkManager', 'openstack-selinux', 'python3-libselinux', 'python3-pyyaml', 'rsync', 'tmpwatch', 'sysstat', 'iproute-tc', 'ksmtuned', 'systemd-container', 'crypto-policies-scripts', 'grubby', 'sos'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14392 DF PROTO=TCP SPT=58830 DPT=9100 SEQ=2774045645 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284D7DA0000000001030307) Nov 27 04:17:32 localhost python3.9[127520]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['openstack-network-scripts'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=50612 DF PROTO=TCP SPT=47204 DPT=9105 SEQ=1024839421 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284E41A0000000001030307) Nov 27 04:17:37 localhost python3.9[127614]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['podman', 'buildah'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59395 DF PROTO=TCP SPT=39920 DPT=9101 SEQ=3991612988 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284F7DB0000000001030307) Nov 27 04:17:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=30847 DF PROTO=TCP SPT=52724 DPT=9882 SEQ=1615907590 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1284FB170000000001030307) Nov 27 04:17:41 localhost python3.9[127714]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['tuned', 'tuned-profiles-cpu-partitioning'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=30849 DF PROTO=TCP SPT=52724 DPT=9882 SEQ=1615907590 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285071A0000000001030307) Nov 27 04:17:45 localhost python3.9[127919]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['os-net-config'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12512 DF PROTO=TCP SPT=39180 DPT=9102 SEQ=1383901047 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128515A90000000001030307) Nov 27 04:17:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63107 DF PROTO=TCP SPT=40178 DPT=9100 SEQ=703172869 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12851DDA0000000001030307) Nov 27 04:17:49 localhost python3.9[128028]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['openssh-server'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 5400.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:17:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63108 DF PROTO=TCP SPT=40178 DPT=9100 SEQ=703172869 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12852D9A0000000001030307) Nov 27 04:17:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29657 DF PROTO=TCP SPT=45608 DPT=9101 SEQ=2046551529 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285359A0000000001030307) Nov 27 04:17:53 localhost python3.9[128122]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['libvirt ', 'libvirt-admin ', 'libvirt-client ', 'libvirt-daemon ', 'qemu-kvm', 'qemu-img', 'libguestfs', 'libseccomp', 'swtpm', 'swtpm-tools', 'edk2-ovmf', 'ceph-common', 'cyrus-sasl-scram'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 5400.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:17:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23157 DF PROTO=TCP SPT=43352 DPT=9105 SEQ=1878074723 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285419B0000000001030307) Nov 27 04:17:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29659 DF PROTO=TCP SPT=45608 DPT=9101 SEQ=2046551529 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12854D5A0000000001030307) Nov 27 04:18:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23159 DF PROTO=TCP SPT=43352 DPT=9105 SEQ=1878074723 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285595A0000000001030307) Nov 27 04:18:06 localhost python3.9[128291]: ansible-ansible.builtin.file Invoked with group=zuul mode=0770 owner=zuul path=/root/.config/containers recurse=True state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:18:06 localhost python3.9[128396]: ansible-ansible.legacy.stat Invoked with path=/root/.config/containers/auth.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:18:07 localhost python3.9[128469]: ansible-ansible.legacy.copy Invoked with dest=/root/.config/containers/auth.json group=zuul mode=0660 owner=zuul src=/home/zuul/.ansible/tmp/ansible-tmp-1764235086.3493001-724-14238447390089/.source.json _original_basename=.7opg2_zf follow=False checksum=bf21a9e8fbc5a3846fb05b4fa0859e0917b2202f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:18:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29660 DF PROTO=TCP SPT=45608 DPT=9101 SEQ=2046551529 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12856DDA0000000001030307) Nov 27 04:18:08 localhost python3.9[128561]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32454 DF PROTO=TCP SPT=35620 DPT=9882 SEQ=864904489 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128570470000000001030307) Nov 27 04:18:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32456 DF PROTO=TCP SPT=35620 DPT=9882 SEQ=864904489 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12857C5A0000000001030307) Nov 27 04:18:14 localhost podman[128574]: 2025-11-27 09:18:08.543081271 +0000 UTC m=+0.045802844 image pull quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified Nov 27 04:18:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13804 DF PROTO=TCP SPT=58906 DPT=9102 SEQ=2803130295 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12858AD90000000001030307) Nov 27 04:18:16 localhost python3.9[128771]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33115 DF PROTO=TCP SPT=38444 DPT=9100 SEQ=2198620232 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285931A0000000001030307) Nov 27 04:18:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33116 DF PROTO=TCP SPT=38444 DPT=9100 SEQ=2198620232 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285A2DB0000000001030307) Nov 27 04:18:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62876 DF PROTO=TCP SPT=42064 DPT=9101 SEQ=816456438 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285AA9A0000000001030307) Nov 27 04:18:24 localhost podman[128783]: 2025-11-27 09:18:16.429802558 +0000 UTC m=+0.045631058 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 04:18:25 localhost python3.9[128982]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1400 DF PROTO=TCP SPT=49474 DPT=9105 SEQ=985599074 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285B6DB0000000001030307) Nov 27 04:18:28 localhost podman[128995]: 2025-11-27 09:18:26.046937871 +0000 UTC m=+0.045325439 image pull quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified Nov 27 04:18:29 localhost python3.9[129161]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=50615 DF PROTO=TCP SPT=47204 DPT=9105 SEQ=1024839421 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285C1DA0000000001030307) Nov 27 04:18:30 localhost podman[129175]: 2025-11-27 09:18:29.614863741 +0000 UTC m=+0.045733830 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 04:18:32 localhost python3.9[129339]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1402 DF PROTO=TCP SPT=49474 DPT=9105 SEQ=985599074 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285CE9A0000000001030307) Nov 27 04:18:35 localhost podman[129351]: 2025-11-27 09:18:32.470790539 +0000 UTC m=+0.046729878 image pull quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified Nov 27 04:18:36 localhost python3.9[129530]: ansible-containers.podman.podman_image Invoked with auth_file=/root/.config/containers/auth.json name=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c tag=latest pull=True push=False force=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'container_file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'ssh': None, 'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'extra_args': None, 'transport': None} arch=None pull_extra_args=None path=None validate_certs=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Nov 27 04:18:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62879 DF PROTO=TCP SPT=42064 DPT=9101 SEQ=816456438 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285E1DA0000000001030307) Nov 27 04:18:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28447 DF PROTO=TCP SPT=56164 DPT=9882 SEQ=2821942065 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285E5770000000001030307) Nov 27 04:18:38 localhost podman[129544]: 2025-11-27 09:18:36.787037059 +0000 UTC m=+0.046181624 image pull quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c Nov 27 04:18:40 localhost systemd[1]: session-40.scope: Deactivated successfully. Nov 27 04:18:40 localhost systemd[1]: session-40.scope: Consumed 1min 29.470s CPU time. Nov 27 04:18:40 localhost systemd-logind[761]: Session 40 logged out. Waiting for processes to exit. Nov 27 04:18:40 localhost systemd-logind[761]: Removed session 40. Nov 27 04:18:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28449 DF PROTO=TCP SPT=56164 DPT=9882 SEQ=2821942065 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1285F19B0000000001030307) Nov 27 04:18:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44354 DF PROTO=TCP SPT=54652 DPT=9102 SEQ=3642560347 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128600090000000001030307) Nov 27 04:18:46 localhost sshd[129657]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:18:46 localhost systemd-logind[761]: New session 41 of user zuul. Nov 27 04:18:46 localhost systemd[1]: Started Session 41 of User zuul. Nov 27 04:18:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11789 DF PROTO=TCP SPT=59764 DPT=9100 SEQ=3753397523 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286081A0000000001030307) Nov 27 04:18:47 localhost python3.9[129914]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:18:49 localhost python3.9[130052]: ansible-ansible.builtin.getent Invoked with database=passwd key=openvswitch fail_key=True service=None split=None Nov 27 04:18:50 localhost python3.9[130196]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:18:50 localhost podman[130237]: Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.401428966 +0000 UTC m=+0.071324910 container create 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, vcs-type=git, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , ceph=True, name=rhceph, release=553, vendor=Red Hat, Inc., RELEASE=main, com.redhat.component=rhceph-container) Nov 27 04:18:50 localhost systemd[1]: Started libpod-conmon-84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132.scope. Nov 27 04:18:50 localhost systemd[1]: Started libcrun container. Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.365275433 +0000 UTC m=+0.035171357 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.474245234 +0000 UTC m=+0.144141148 container init 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, CEPH_POINT_RELEASE=, ceph=True, name=rhceph, version=7, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , GIT_BRANCH=main, GIT_CLEAN=True, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, distribution-scope=public) Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.485249848 +0000 UTC m=+0.155145782 container start 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, vcs-type=git, name=rhceph, ceph=True, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, RELEASE=main, CEPH_POINT_RELEASE=, architecture=x86_64, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , version=7, build-date=2025-09-24T08:57:55) Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.48564125 +0000 UTC m=+0.155537194 container attach 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , architecture=x86_64, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, name=rhceph, vcs-type=git, RELEASE=main, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, GIT_BRANCH=main, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., distribution-scope=public, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:18:50 localhost romantic_ride[130258]: 167 167 Nov 27 04:18:50 localhost systemd[1]: libpod-84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132.scope: Deactivated successfully. Nov 27 04:18:50 localhost podman[130237]: 2025-11-27 09:18:50.488449385 +0000 UTC m=+0.158345369 container died 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, release=553, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, build-date=2025-09-24T08:57:55, version=7, distribution-scope=public, architecture=x86_64, io.openshift.expose-services=, maintainer=Guillaume Abrioux , GIT_BRANCH=main, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:18:50 localhost podman[130263]: 2025-11-27 09:18:50.626681402 +0000 UTC m=+0.126484032 container remove 84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_ride, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, GIT_BRANCH=main, architecture=x86_64, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.buildah.version=1.33.12, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, CEPH_POINT_RELEASE=, GIT_CLEAN=True, RELEASE=main) Nov 27 04:18:50 localhost systemd[1]: libpod-conmon-84aded9ae77298752583c6154af2d529f953ab790796496cc34ef32f4ce24132.scope: Deactivated successfully. Nov 27 04:18:50 localhost podman[130287]: Nov 27 04:18:50 localhost podman[130287]: 2025-11-27 09:18:50.845621178 +0000 UTC m=+0.082433757 container create 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, io.openshift.tags=rhceph ceph, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, ceph=True, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, GIT_CLEAN=True, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553) Nov 27 04:18:50 localhost systemd[1]: Started libpod-conmon-4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a.scope. Nov 27 04:18:50 localhost systemd[1]: Started libcrun container. Nov 27 04:18:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/fc6791e4a8b065784f79e3f03247138d1cb07827ac6f0b21c17ebee7aad82b4a/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 04:18:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/fc6791e4a8b065784f79e3f03247138d1cb07827ac6f0b21c17ebee7aad82b4a/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:18:50 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/fc6791e4a8b065784f79e3f03247138d1cb07827ac6f0b21c17ebee7aad82b4a/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:18:50 localhost podman[130287]: 2025-11-27 09:18:50.899334852 +0000 UTC m=+0.136147441 container init 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, io.buildah.version=1.33.12, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, release=553, architecture=x86_64, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_CLEAN=True, ceph=True, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux ) Nov 27 04:18:50 localhost podman[130287]: 2025-11-27 09:18:50.906922766 +0000 UTC m=+0.143735325 container start 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, CEPH_POINT_RELEASE=, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., distribution-scope=public, maintainer=Guillaume Abrioux , GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, ceph=True, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, com.redhat.component=rhceph-container, vcs-type=git, name=rhceph, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:18:50 localhost podman[130287]: 2025-11-27 09:18:50.907079411 +0000 UTC m=+0.143892000 container attach 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, version=7, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, GIT_CLEAN=True, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, vendor=Red Hat, Inc.) Nov 27 04:18:50 localhost podman[130287]: 2025-11-27 09:18:50.811223103 +0000 UTC m=+0.048035722 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:18:51 localhost python3.9[130354]: ansible-ansible.legacy.dnf Invoked with download_only=True name=['openvswitch3.3'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:18:51 localhost systemd[1]: var-lib-containers-storage-overlay-f4007a9458413103f39c3467b9e3d5de51b45921053437fac662c598458cc868-merged.mount: Deactivated successfully. Nov 27 04:18:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11790 DF PROTO=TCP SPT=59764 DPT=9100 SEQ=3753397523 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128617DA0000000001030307) Nov 27 04:18:51 localhost bold_poitras[130330]: [ Nov 27 04:18:51 localhost bold_poitras[130330]: { Nov 27 04:18:51 localhost bold_poitras[130330]: "available": false, Nov 27 04:18:51 localhost bold_poitras[130330]: "ceph_device": false, Nov 27 04:18:51 localhost bold_poitras[130330]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 04:18:51 localhost bold_poitras[130330]: "lsm_data": {}, Nov 27 04:18:51 localhost bold_poitras[130330]: "lvs": [], Nov 27 04:18:51 localhost bold_poitras[130330]: "path": "/dev/sr0", Nov 27 04:18:51 localhost bold_poitras[130330]: "rejected_reasons": [ Nov 27 04:18:51 localhost bold_poitras[130330]: "Insufficient space (<5GB)", Nov 27 04:18:51 localhost bold_poitras[130330]: "Has a FileSystem" Nov 27 04:18:51 localhost bold_poitras[130330]: ], Nov 27 04:18:51 localhost bold_poitras[130330]: "sys_api": { Nov 27 04:18:51 localhost bold_poitras[130330]: "actuators": null, Nov 27 04:18:51 localhost bold_poitras[130330]: "device_nodes": "sr0", Nov 27 04:18:51 localhost bold_poitras[130330]: "human_readable_size": "482.00 KB", Nov 27 04:18:51 localhost bold_poitras[130330]: "id_bus": "ata", Nov 27 04:18:51 localhost bold_poitras[130330]: "model": "QEMU DVD-ROM", Nov 27 04:18:51 localhost bold_poitras[130330]: "nr_requests": "2", Nov 27 04:18:51 localhost bold_poitras[130330]: "partitions": {}, Nov 27 04:18:51 localhost bold_poitras[130330]: "path": "/dev/sr0", Nov 27 04:18:51 localhost bold_poitras[130330]: "removable": "1", Nov 27 04:18:51 localhost bold_poitras[130330]: "rev": "2.5+", Nov 27 04:18:51 localhost bold_poitras[130330]: "ro": "0", Nov 27 04:18:51 localhost bold_poitras[130330]: "rotational": "1", Nov 27 04:18:51 localhost bold_poitras[130330]: "sas_address": "", Nov 27 04:18:51 localhost bold_poitras[130330]: "sas_device_handle": "", Nov 27 04:18:51 localhost bold_poitras[130330]: "scheduler_mode": "mq-deadline", Nov 27 04:18:51 localhost bold_poitras[130330]: "sectors": 0, Nov 27 04:18:51 localhost bold_poitras[130330]: "sectorsize": "2048", Nov 27 04:18:51 localhost bold_poitras[130330]: "size": 493568.0, Nov 27 04:18:51 localhost bold_poitras[130330]: "support_discard": "0", Nov 27 04:18:51 localhost bold_poitras[130330]: "type": "disk", Nov 27 04:18:51 localhost bold_poitras[130330]: "vendor": "QEMU" Nov 27 04:18:51 localhost bold_poitras[130330]: } Nov 27 04:18:51 localhost bold_poitras[130330]: } Nov 27 04:18:51 localhost bold_poitras[130330]: ] Nov 27 04:18:51 localhost systemd[1]: libpod-4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a.scope: Deactivated successfully. Nov 27 04:18:51 localhost podman[130287]: 2025-11-27 09:18:51.694407669 +0000 UTC m=+0.931220238 container died 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, CEPH_POINT_RELEASE=, vcs-type=git, release=553, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, vendor=Red Hat, Inc., GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, name=rhceph, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, maintainer=Guillaume Abrioux ) Nov 27 04:18:51 localhost systemd[1]: tmp-crun.BWHSJc.mount: Deactivated successfully. Nov 27 04:18:51 localhost systemd[1]: var-lib-containers-storage-overlay-fc6791e4a8b065784f79e3f03247138d1cb07827ac6f0b21c17ebee7aad82b4a-merged.mount: Deactivated successfully. Nov 27 04:18:51 localhost podman[131762]: 2025-11-27 09:18:51.81492885 +0000 UTC m=+0.113055811 container remove 4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=bold_poitras, io.k8s.description=Red Hat Ceph Storage 7, version=7, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, architecture=x86_64, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_BRANCH=main, name=rhceph, build-date=2025-09-24T08:57:55, vcs-type=git, RELEASE=main, io.openshift.expose-services=, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:18:51 localhost systemd[1]: libpod-conmon-4c9e2206bfc7590172abcbdd872f0cc6f87fb9e1712035d0fa170874e1a97b7a.scope: Deactivated successfully. Nov 27 04:18:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64200 DF PROTO=TCP SPT=41782 DPT=9101 SEQ=475424956 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12861FDB0000000001030307) Nov 27 04:18:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=29662 DF PROTO=TCP SPT=45608 DPT=9101 SEQ=2046551529 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12862BDB0000000001030307) Nov 27 04:18:57 localhost python3.9[132297]: ansible-ansible.legacy.dnf Invoked with name=['openvswitch3.3'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:18:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64202 DF PROTO=TCP SPT=41782 DPT=9101 SEQ=475424956 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286379A0000000001030307) Nov 27 04:19:02 localhost python3.9[132391]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=openvswitch.service state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:19:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1794 DF PROTO=TCP SPT=43824 DPT=9105 SEQ=2403709680 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128643DA0000000001030307) Nov 27 04:19:03 localhost python3.9[132484]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'selinux'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:19:04 localhost python3.9[132576]: ansible-community.general.sefcontext Invoked with selevel=s0 setype=container_file_t state=present target=/var/lib/edpm-config(/.*)? ignore_selinux_state=False ftype=a reload=True substitute=None seuser=None Nov 27 04:19:07 localhost kernel: SELinux: Converting 2757 SID table entries... Nov 27 04:19:07 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:19:07 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:19:07 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:19:07 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:19:07 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:19:07 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:19:07 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:19:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64203 DF PROTO=TCP SPT=41782 DPT=9101 SEQ=475424956 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128657DA0000000001030307) Nov 27 04:19:08 localhost python3.9[133167]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local', 'distribution'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:19:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64440 DF PROTO=TCP SPT=55560 DPT=9882 SEQ=3254679663 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12865AA70000000001030307) Nov 27 04:19:08 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=18 res=1 Nov 27 04:19:09 localhost python3.9[133265]: ansible-ansible.legacy.dnf Invoked with name=['driverctl', 'lvm2', 'crudini', 'jq', 'nftables', 'NetworkManager', 'openstack-selinux', 'python3-libselinux', 'python3-pyyaml', 'rsync', 'tmpwatch', 'sysstat', 'iproute-tc', 'ksmtuned', 'systemd-container', 'crypto-policies-scripts', 'grubby', 'sos'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:19:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64442 DF PROTO=TCP SPT=55560 DPT=9882 SEQ=3254679663 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286669A0000000001030307) Nov 27 04:19:13 localhost python3.9[133359]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -V driverctl lvm2 crudini jq nftables NetworkManager openstack-selinux python3-libselinux python3-pyyaml rsync tmpwatch sysstat iproute-tc ksmtuned systemd-container crypto-policies-scripts grubby sos _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:19:14 localhost python3.9[133604]: ansible-ansible.builtin.file Invoked with mode=0750 path=/var/lib/edpm-config selevel=s0 setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 04:19:15 localhost python3.9[133694]: ansible-ansible.builtin.stat Invoked with path=/etc/cloud/cloud.cfg.d follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:19:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64748 DF PROTO=TCP SPT=39252 DPT=9102 SEQ=2514837814 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286753A0000000001030307) Nov 27 04:19:16 localhost python3.9[133788]: ansible-ansible.legacy.dnf Invoked with name=['os-net-config'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:19:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47884 DF PROTO=TCP SPT=34672 DPT=9100 SEQ=1681147372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12867D5A0000000001030307) Nov 27 04:19:20 localhost python3.9[133882]: ansible-ansible.legacy.dnf Invoked with name=['openstack-network-scripts'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:19:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47885 DF PROTO=TCP SPT=34672 DPT=9100 SEQ=1681147372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12868D1A0000000001030307) Nov 27 04:19:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54674 DF PROTO=TCP SPT=35338 DPT=9101 SEQ=692054903 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286951A0000000001030307) Nov 27 04:19:24 localhost python3.9[133976]: ansible-ansible.builtin.systemd Invoked with enabled=True name=network daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 04:19:24 localhost systemd[1]: Reloading. Nov 27 04:19:24 localhost systemd-rc-local-generator[134003]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:19:24 localhost systemd-sysv-generator[134008]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:19:24 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:19:25 localhost python3.9[134108]: ansible-ansible.builtin.stat Invoked with path=/var/lib/edpm-config/os-net-config.returncode follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:19:26 localhost python3.9[134200]: ansible-community.general.ini_file Invoked with backup=True mode=0644 no_extra_spaces=True option=no-auto-default path=/etc/NetworkManager/NetworkManager.conf section=main state=present value=* exclusive=True ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False section_has_values=None values=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6552 DF PROTO=TCP SPT=39736 DPT=9105 SEQ=3210869602 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286A11A0000000001030307) Nov 27 04:19:27 localhost python3.9[134294]: ansible-community.general.ini_file Invoked with backup=True mode=0644 no_extra_spaces=True option=dns path=/etc/NetworkManager/NetworkManager.conf section=main state=present value=none exclusive=True ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False section_has_values=None values=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:28 localhost python3.9[134386]: ansible-community.general.ini_file Invoked with backup=True mode=0644 no_extra_spaces=True option=rc-manager path=/etc/NetworkManager/NetworkManager.conf section=main state=present value=unmanaged exclusive=True ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False section_has_values=None values=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54676 DF PROTO=TCP SPT=35338 DPT=9101 SEQ=692054903 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286ACDA0000000001030307) Nov 27 04:19:29 localhost python3.9[134478]: ansible-ansible.legacy.stat Invoked with path=/etc/dhcp/dhclient-enter-hooks follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:19:30 localhost python3.9[134551]: ansible-ansible.legacy.copy Invoked with dest=/etc/dhcp/dhclient-enter-hooks mode=0755 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235169.3878658-565-198955586708746/.source _original_basename=.004r9712 follow=False checksum=f6278a40de79a9841f6ed1fc584538225566990c backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:31 localhost python3.9[134643]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/os-net-config state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:32 localhost python3.9[134735]: ansible-edpm_os_net_config_mappings Invoked with net_config_data_lookup={} Nov 27 04:19:32 localhost python3.9[134827]: ansible-ansible.builtin.file Invoked with path=/var/lib/edpm-config/scripts state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6554 DF PROTO=TCP SPT=39736 DPT=9105 SEQ=3210869602 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286B8DC0000000001030307) Nov 27 04:19:33 localhost python3.9[134919]: ansible-ansible.legacy.stat Invoked with path=/etc/os-net-config/config.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:19:34 localhost python3.9[134992]: ansible-ansible.legacy.copy Invoked with backup=True dest=/etc/os-net-config/config.yaml mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235173.1221955-691-37859150942985/.source.yaml _original_basename=.qbdgyvyj follow=False checksum=4c28d1662755c608a6ffaa942e27a2488c0a78a3 force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:35 localhost python3.9[135084]: ansible-ansible.builtin.slurp Invoked with path=/etc/os-net-config/config.yaml src=/etc/os-net-config/config.yaml Nov 27 04:19:36 localhost ansible-async_wrapper.py[135189]: Invoked with j768132681389 300 /home/zuul/.ansible/tmp/ansible-tmp-1764235175.4710069-763-179458041434765/AnsiballZ_edpm_os_net_config.py _ Nov 27 04:19:36 localhost ansible-async_wrapper.py[135192]: Starting module and watcher Nov 27 04:19:36 localhost ansible-async_wrapper.py[135192]: Start watching 135193 (300) Nov 27 04:19:36 localhost ansible-async_wrapper.py[135193]: Start module (135193) Nov 27 04:19:36 localhost ansible-async_wrapper.py[135189]: Return async_wrapper task started. Nov 27 04:19:36 localhost python3.9[135194]: ansible-edpm_os_net_config Invoked with cleanup=False config_file=/etc/os-net-config/config.yaml debug=True detailed_exit_codes=True safe_defaults=False use_nmstate=False Nov 27 04:19:37 localhost ansible-async_wrapper.py[135193]: Module complete (135193) Nov 27 04:19:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=54677 DF PROTO=TCP SPT=35338 DPT=9101 SEQ=692054903 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286CDDA0000000001030307) Nov 27 04:19:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41130 DF PROTO=TCP SPT=35478 DPT=9882 SEQ=2552857824 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286CFD70000000001030307) Nov 27 04:19:39 localhost python3.9[135286]: ansible-ansible.legacy.async_status Invoked with jid=j768132681389.135189 mode=status _async_dir=/root/.ansible_async Nov 27 04:19:40 localhost python3.9[135345]: ansible-ansible.legacy.async_status Invoked with jid=j768132681389.135189 mode=cleanup _async_dir=/root/.ansible_async Nov 27 04:19:41 localhost ansible-async_wrapper.py[135192]: Done in kid B. Nov 27 04:19:41 localhost python3.9[135437]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/os-net-config.returncode follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:19:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41132 DF PROTO=TCP SPT=35478 DPT=9882 SEQ=2552857824 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286DBDA0000000001030307) Nov 27 04:19:42 localhost python3.9[135510]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/edpm-config/os-net-config.returncode mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235181.321049-829-20545944362978/.source.returncode _original_basename=.t3880ybv follow=False checksum=b6589fc6ab0dc82cf12099d1c2d40ab994e8410c backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:42 localhost python3.9[135602]: ansible-ansible.legacy.stat Invoked with path=/etc/cloud/cloud.cfg.d/99-edpm-disable-network-config.cfg follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:19:43 localhost python3.9[135675]: ansible-ansible.legacy.copy Invoked with dest=/etc/cloud/cloud.cfg.d/99-edpm-disable-network-config.cfg mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235182.5362768-877-34707994287009/.source.cfg _original_basename=.la5jdhw9 follow=False checksum=f3c5952a9cd4c6c31b314b25eb897168971cc86e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:19:44 localhost python3.9[135767]: ansible-ansible.builtin.systemd Invoked with name=NetworkManager state=reloaded daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:19:44 localhost systemd[1]: Reloading Network Manager... Nov 27 04:19:44 localhost NetworkManager[5971]: [1764235184.3716] audit: op="reload" arg="0" pid=135771 uid=0 result="success" Nov 27 04:19:44 localhost NetworkManager[5971]: [1764235184.3727] config: signal: SIGHUP (no changes from disk) Nov 27 04:19:44 localhost systemd[1]: Reloaded Network Manager. Nov 27 04:19:44 localhost systemd[1]: session-41.scope: Deactivated successfully. Nov 27 04:19:44 localhost systemd[1]: session-41.scope: Consumed 35.506s CPU time. Nov 27 04:19:44 localhost systemd-logind[761]: Session 41 logged out. Waiting for processes to exit. Nov 27 04:19:44 localhost systemd-logind[761]: Removed session 41. Nov 27 04:19:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33944 DF PROTO=TCP SPT=41832 DPT=9102 SEQ=1832225597 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286EA6B0000000001030307) Nov 27 04:19:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13874 DF PROTO=TCP SPT=49644 DPT=9100 SEQ=2606852239 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1286F29A0000000001030307) Nov 27 04:19:50 localhost sshd[135786]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:19:50 localhost systemd-logind[761]: New session 42 of user zuul. Nov 27 04:19:50 localhost systemd[1]: Started Session 42 of User zuul. Nov 27 04:19:51 localhost python3.9[135879]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:19:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13875 DF PROTO=TCP SPT=49644 DPT=9100 SEQ=2606852239 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287025A0000000001030307) Nov 27 04:19:53 localhost python3.9[136003]: ansible-ansible.builtin.setup Invoked with filter=['ansible_default_ipv4'] gather_subset=['!all', '!min', 'network'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:19:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6680 DF PROTO=TCP SPT=39340 DPT=9101 SEQ=2412076699 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12870A5A0000000001030307) Nov 27 04:19:54 localhost python3.9[136189]: ansible-ansible.legacy.command Invoked with _raw_params=hostname -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:19:54 localhost systemd-logind[761]: Session 42 logged out. Waiting for processes to exit. Nov 27 04:19:54 localhost systemd[1]: session-42.scope: Deactivated successfully. Nov 27 04:19:54 localhost systemd[1]: session-42.scope: Consumed 2.058s CPU time. Nov 27 04:19:54 localhost systemd-logind[761]: Removed session 42. Nov 27 04:19:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64205 DF PROTO=TCP SPT=41782 DPT=9101 SEQ=475424956 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128715DB0000000001030307) Nov 27 04:19:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13876 DF PROTO=TCP SPT=49644 DPT=9100 SEQ=2606852239 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128721DA0000000001030307) Nov 27 04:20:00 localhost sshd[136220]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:20:00 localhost systemd-logind[761]: New session 43 of user zuul. Nov 27 04:20:00 localhost systemd[1]: Started Session 43 of User zuul. Nov 27 04:20:01 localhost python3.9[136313]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:20:02 localhost python3.9[136407]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:20:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49453 DF PROTO=TCP SPT=45762 DPT=9105 SEQ=2854117071 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12872E1B0000000001030307) Nov 27 04:20:03 localhost python3.9[136503]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:20:04 localhost python3.9[136557]: ansible-ansible.legacy.dnf Invoked with name=['podman'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:20:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6683 DF PROTO=TCP SPT=39340 DPT=9101 SEQ=2412076699 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128741DA0000000001030307) Nov 27 04:20:08 localhost python3.9[136651]: ansible-ansible.builtin.setup Invoked with filter=['ansible_interfaces'] gather_subset=['!all', '!min', 'network'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:20:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32934 DF PROTO=TCP SPT=40938 DPT=9882 SEQ=3924622119 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128745060000000001030307) Nov 27 04:20:10 localhost python3.9[136806]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/containers/networks recurse=True state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:20:10 localhost python3.9[136898]: ansible-ansible.legacy.command Invoked with _raw_params=podman network inspect podman#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:20:11 localhost python3.9[137001]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/networks/podman.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:20:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32936 DF PROTO=TCP SPT=40938 DPT=9882 SEQ=3924622119 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287511A0000000001030307) Nov 27 04:20:12 localhost python3.9[137049]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/containers/networks/podman.json _original_basename=podman_network_config.j2 recurse=False state=file path=/etc/containers/networks/podman.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:20:12 localhost python3.9[137141]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/20-edpm-podman-registries.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:20:13 localhost python3.9[137189]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root setype=etc_t dest=/etc/containers/registries.conf.d/20-edpm-podman-registries.conf _original_basename=registries.conf.j2 recurse=False state=file path=/etc/containers/registries.conf.d/20-edpm-podman-registries.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:20:14 localhost python3.9[137281]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=pids_limit owner=root path=/etc/containers/containers.conf section=containers setype=etc_t value=4096 backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:20:14 localhost python3.9[137373]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=events_logger owner=root path=/etc/containers/containers.conf section=engine setype=etc_t value="journald" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:20:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13444 DF PROTO=TCP SPT=56096 DPT=9102 SEQ=3642238324 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12875F990000000001030307) Nov 27 04:20:15 localhost python3.9[137465]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=runtime owner=root path=/etc/containers/containers.conf section=engine setype=etc_t value="crun" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:20:16 localhost python3.9[137557]: ansible-community.general.ini_file Invoked with create=True group=root mode=0644 option=network_backend owner=root path=/etc/containers/containers.conf section=network setype=etc_t value="netavark" backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True follow=False unsafe_writes=False section_has_values=None values=None seuser=None serole=None selevel=None attributes=None Nov 27 04:20:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37057 DF PROTO=TCP SPT=38300 DPT=9100 SEQ=529871197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128767DA0000000001030307) Nov 27 04:20:17 localhost python3.9[137649]: ansible-ansible.legacy.dnf Invoked with name=['openssh-server'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:20:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37058 DF PROTO=TCP SPT=38300 DPT=9100 SEQ=529871197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287779B0000000001030307) Nov 27 04:20:22 localhost python3.9[137743]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:20:22 localhost python3.9[137837]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:20:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8565 DF PROTO=TCP SPT=44088 DPT=9101 SEQ=2481805036 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12877F5A0000000001030307) Nov 27 04:20:23 localhost python3.9[137929]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:20:24 localhost python3.9[138021]: ansible-ansible.legacy.command Invoked with _raw_params=systemctl is-system-running _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:20:25 localhost python3.9[138114]: ansible-service_facts Invoked Nov 27 04:20:25 localhost network[138131]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:20:25 localhost network[138132]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:20:25 localhost network[138133]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:20:26 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:20:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21294 DF PROTO=TCP SPT=58484 DPT=9105 SEQ=445038921 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12878B9A0000000001030307) Nov 27 04:20:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8567 DF PROTO=TCP SPT=44088 DPT=9101 SEQ=2481805036 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287971B0000000001030307) Nov 27 04:20:32 localhost python3.9[138454]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:20:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21296 DF PROTO=TCP SPT=58484 DPT=9105 SEQ=445038921 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287A35A0000000001030307) Nov 27 04:20:37 localhost python3.9[138548]: ansible-package_facts Invoked with manager=['auto'] strategy=first Nov 27 04:20:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8568 DF PROTO=TCP SPT=44088 DPT=9101 SEQ=2481805036 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287B7DA0000000001030307) Nov 27 04:20:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57742 DF PROTO=TCP SPT=33270 DPT=9882 SEQ=3887614405 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287BA370000000001030307) Nov 27 04:20:40 localhost python3.9[138640]: ansible-ansible.legacy.stat Invoked with path=/etc/chrony.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:20:41 localhost python3.9[138715]: ansible-ansible.legacy.copy Invoked with backup=True dest=/etc/chrony.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235238.7717419-659-134508589473161/.source.conf follow=False _original_basename=chrony.conf.j2 checksum=cfb003e56d02d0d2c65555452eb1a05073fecdad force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:20:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57744 DF PROTO=TCP SPT=33270 DPT=9882 SEQ=3887614405 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287C65A0000000001030307) Nov 27 04:20:41 localhost python3.9[138809]: ansible-ansible.legacy.stat Invoked with path=/etc/sysconfig/chronyd follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:20:42 localhost python3.9[138884]: ansible-ansible.legacy.copy Invoked with backup=True dest=/etc/sysconfig/chronyd mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235241.392358-705-97033076841070/.source follow=False _original_basename=chronyd.sysconfig.j2 checksum=dd196b1ff1f915b23eebc37ec77405b5dd3df76c force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:20:44 localhost python3.9[138978]: ansible-lineinfile Invoked with backup=True create=True dest=/etc/sysconfig/network line=PEERNTP=no mode=0644 regexp=^PEERNTP= state=present path=/etc/sysconfig/network encoding=utf-8 backrefs=False firstmatch=False unsafe_writes=False search_string=None insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:20:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21815 DF PROTO=TCP SPT=43250 DPT=9102 SEQ=3815685174 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287D4C80000000001030307) Nov 27 04:20:46 localhost python3.9[139072]: ansible-ansible.legacy.setup Invoked with gather_subset=['!all'] filter=['ansible_service_mgr'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:20:47 localhost python3.9[139126]: ansible-ansible.legacy.systemd Invoked with enabled=True name=chronyd state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:20:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62394 DF PROTO=TCP SPT=49600 DPT=9100 SEQ=3728545456 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287DCDA0000000001030307) Nov 27 04:20:49 localhost python3.9[139220]: ansible-ansible.legacy.setup Invoked with gather_subset=['!all'] filter=['ansible_service_mgr'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:20:50 localhost python3.9[139274]: ansible-ansible.legacy.systemd Invoked with name=chronyd state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:20:50 localhost chronyd[25955]: chronyd exiting Nov 27 04:20:50 localhost systemd[1]: Stopping NTP client/server... Nov 27 04:20:50 localhost systemd[1]: chronyd.service: Deactivated successfully. Nov 27 04:20:50 localhost systemd[1]: Stopped NTP client/server. Nov 27 04:20:50 localhost systemd[1]: Starting NTP client/server... Nov 27 04:20:50 localhost chronyd[139282]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Nov 27 04:20:50 localhost chronyd[139282]: Frequency -26.541 +/- 0.241 ppm read from /var/lib/chrony/drift Nov 27 04:20:50 localhost chronyd[139282]: Loaded seccomp filter (level 2) Nov 27 04:20:50 localhost systemd[1]: Started NTP client/server. Nov 27 04:20:51 localhost systemd[1]: session-43.scope: Deactivated successfully. Nov 27 04:20:51 localhost systemd[1]: session-43.scope: Consumed 28.077s CPU time. Nov 27 04:20:51 localhost systemd-logind[761]: Session 43 logged out. Waiting for processes to exit. Nov 27 04:20:51 localhost systemd-logind[761]: Removed session 43. Nov 27 04:20:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62395 DF PROTO=TCP SPT=49600 DPT=9100 SEQ=3728545456 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287EC9A0000000001030307) Nov 27 04:20:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11753 DF PROTO=TCP SPT=49426 DPT=9101 SEQ=2454803445 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287F49A0000000001030307) Nov 27 04:20:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6685 DF PROTO=TCP SPT=39340 DPT=9101 SEQ=2412076699 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1287FFDA0000000001030307) Nov 27 04:20:57 localhost sshd[139397]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:20:57 localhost systemd-logind[761]: New session 44 of user zuul. Nov 27 04:20:57 localhost systemd[1]: Started Session 44 of User zuul. Nov 27 04:20:58 localhost python3.9[139519]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:20:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49456 DF PROTO=TCP SPT=45762 DPT=9105 SEQ=2854117071 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12880BDA0000000001030307) Nov 27 04:20:59 localhost python3.9[139615]: ansible-ansible.builtin.file Invoked with group=zuul mode=0770 owner=zuul path=/root/.config/containers recurse=True state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:00 localhost python3.9[139720]: ansible-ansible.legacy.stat Invoked with path=/root/.config/containers/auth.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:01 localhost python3.9[139768]: ansible-ansible.legacy.file Invoked with group=zuul mode=0660 owner=zuul dest=/root/.config/containers/auth.json _original_basename=.djn00l5a recurse=False state=file path=/root/.config/containers/auth.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:02 localhost python3.9[139860]: ansible-ansible.legacy.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41843 DF PROTO=TCP SPT=35360 DPT=9105 SEQ=1657358675 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288189A0000000001030307) Nov 27 04:21:02 localhost python3.9[139935]: ansible-ansible.legacy.copy Invoked with dest=/etc/sysconfig/podman_drop_in mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235261.7097442-145-20405286962517/.source _original_basename=.fh7iaegd follow=False checksum=125299ce8dea7711a76292961206447f0043248b backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:03 localhost python3.9[140027]: ansible-ansible.builtin.file Invoked with path=/var/local/libexec recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:21:04 localhost python3.9[140119]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-container-shutdown follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:04 localhost python3.9[140192]: ansible-ansible.legacy.copy Invoked with dest=/var/local/libexec/edpm-container-shutdown group=root mode=0700 owner=root setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235263.8874876-218-124429594963554/.source _original_basename=edpm-container-shutdown follow=False checksum=632c3792eb3dce4288b33ae7b265b71950d69f13 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:21:04 localhost auditd[715]: Audit daemon rotating log files Nov 27 04:21:05 localhost python3.9[140284]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-start-podman-container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:06 localhost python3.9[140357]: ansible-ansible.legacy.copy Invoked with dest=/var/local/libexec/edpm-start-podman-container group=root mode=0700 owner=root setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235265.0522954-218-105552340686590/.source _original_basename=edpm-start-podman-container follow=False checksum=b963c569d75a655c0ccae95d9bb4a2a9a4df27d1 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:21:06 localhost python3.9[140449]: ansible-ansible.builtin.file Invoked with mode=420 path=/etc/systemd/system-preset state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:07 localhost python3.9[140541]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm-container-shutdown.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11756 DF PROTO=TCP SPT=49426 DPT=9101 SEQ=2454803445 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12882BDA0000000001030307) Nov 27 04:21:08 localhost python3.9[140614]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/edpm-container-shutdown.service group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235266.8911815-328-8734410844974/.source.service _original_basename=edpm-container-shutdown-service follow=False checksum=6336835cb0f888670cc99de31e19c8c071444d33 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37427 DF PROTO=TCP SPT=34984 DPT=9882 SEQ=81233788 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12882F660000000001030307) Nov 27 04:21:08 localhost python3.9[140706]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:09 localhost python3.9[140779]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system-preset/91-edpm-container-shutdown.preset group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235268.3513854-373-138364457353429/.source.preset _original_basename=91-edpm-container-shutdown-preset follow=False checksum=b275e4375287528cb63464dd32f622c4f142a915 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:10 localhost python3.9[140871]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm-container-shutdown state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:21:10 localhost systemd[1]: Reloading. Nov 27 04:21:10 localhost systemd-rc-local-generator[140898]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:21:10 localhost systemd-sysv-generator[140901]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:21:10 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:21:10 localhost systemd[1]: Reloading. Nov 27 04:21:11 localhost systemd-rc-local-generator[140931]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:21:11 localhost systemd-sysv-generator[140934]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:21:11 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:21:11 localhost systemd[1]: Starting EDPM Container Shutdown... Nov 27 04:21:11 localhost systemd[1]: Finished EDPM Container Shutdown. Nov 27 04:21:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37429 DF PROTO=TCP SPT=34984 DPT=9882 SEQ=81233788 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12883B5B0000000001030307) Nov 27 04:21:11 localhost python3.9[141038]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/netns-placeholder.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:12 localhost python3.9[141111]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/netns-placeholder.service group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235271.4860768-442-207777829397843/.source.service _original_basename=netns-placeholder-service follow=False checksum=b61b1b5918c20c877b8b226fbf34ff89a082d972 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:13 localhost python3.9[141203]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:13 localhost python3.9[141276]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system-preset/91-netns-placeholder.preset group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235272.7246718-487-173114488628288/.source.preset _original_basename=91-netns-placeholder-preset follow=False checksum=28b7b9aa893525d134a1eeda8a0a48fb25b736b9 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:14 localhost python3.9[141368]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=netns-placeholder state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:21:14 localhost systemd[1]: Reloading. Nov 27 04:21:14 localhost systemd-rc-local-generator[141396]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:21:14 localhost systemd-sysv-generator[141399]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:21:14 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:21:15 localhost systemd[1]: Starting Create netns directory... Nov 27 04:21:15 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 04:21:15 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 04:21:15 localhost systemd[1]: Finished Create netns directory. Nov 27 04:21:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4622 DF PROTO=TCP SPT=43806 DPT=9102 SEQ=4271669185 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128849F90000000001030307) Nov 27 04:21:16 localhost python3.9[141500]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:21:16 localhost network[141517]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:21:16 localhost network[141518]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:21:16 localhost network[141519]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:21:17 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:21:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36161 DF PROTO=TCP SPT=60148 DPT=9100 SEQ=1707789345 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288521A0000000001030307) Nov 27 04:21:21 localhost python3.9[141720]: ansible-ansible.legacy.stat Invoked with path=/etc/ssh/sshd_config follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36162 DF PROTO=TCP SPT=60148 DPT=9100 SEQ=1707789345 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128861DB0000000001030307) Nov 27 04:21:22 localhost python3.9[141795]: ansible-ansible.legacy.copy Invoked with dest=/etc/ssh/sshd_config mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235280.4728599-611-54210395108190/.source validate=/usr/sbin/sshd -T -f %s follow=False _original_basename=sshd_config_block.j2 checksum=6c79f4cb960ad444688fde322eeacb8402e22d79 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9043 DF PROTO=TCP SPT=49984 DPT=9101 SEQ=262484235 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128869DA0000000001030307) Nov 27 04:21:24 localhost python3.9[141888]: ansible-ansible.builtin.systemd Invoked with name=sshd state=reloaded daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:21:24 localhost systemd[1]: Reloading OpenSSH server daemon... Nov 27 04:21:24 localhost systemd[1]: Reloaded OpenSSH server daemon. Nov 27 04:21:24 localhost sshd[121241]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:21:25 localhost python3.9[141984]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/var/lib/edpm-config/firewall state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:25 localhost python3.9[142076]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/sshd-networks.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:26 localhost python3.9[142149]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/edpm-config/firewall/sshd-networks.yaml group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235285.4294565-705-171686998502310/.source.yaml follow=False _original_basename=firewall.yaml.j2 checksum=0bfc8440fd8f39002ab90252479fb794f51b5ae8 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40454 DF PROTO=TCP SPT=36408 DPT=9105 SEQ=394258578 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128875DA0000000001030307) Nov 27 04:21:27 localhost python3.9[142241]: ansible-community.general.timezone Invoked with name=UTC hwclock=None Nov 27 04:21:27 localhost systemd[1]: Starting Time & Date Service... Nov 27 04:21:27 localhost systemd[1]: Started Time & Date Service. Nov 27 04:21:28 localhost python3.9[142337]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/var/lib/edpm-config/firewall state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:29 localhost python3.9[142429]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9045 DF PROTO=TCP SPT=49984 DPT=9101 SEQ=262484235 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288819A0000000001030307) Nov 27 04:21:29 localhost python3.9[142502]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235288.7630847-809-179246385404096/.source.yaml follow=False _original_basename=base-rules.yaml.j2 checksum=450456afcafded6d4bdecceec7a02e806eebd8b3 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:30 localhost python3.9[142594]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:31 localhost python3.9[142667]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235290.3403826-854-138991008279486/.source.yaml _original_basename=.21t0phtb follow=False checksum=97d170e1550eee4afc0af065b78cda302a97674c backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:31 localhost python3.9[142759]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/iptables.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40456 DF PROTO=TCP SPT=36408 DPT=9105 SEQ=394258578 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12888D9A0000000001030307) Nov 27 04:21:33 localhost python3.9[142834]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/iptables.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235291.535444-899-271428975178846/.source.nft _original_basename=iptables.nft follow=False checksum=3e02df08f1f3ab4a513e94056dbd390e3d38fe30 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:34 localhost python3.9[142926]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/iptables.nft _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:21:35 localhost python3.9[143019]: ansible-ansible.legacy.command Invoked with _raw_params=nft -j list ruleset _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:21:36 localhost python3[143112]: ansible-edpm_nftables_from_files Invoked with src=/var/lib/edpm-config/firewall Nov 27 04:21:36 localhost python3.9[143204]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:37 localhost python3.9[143277]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-jumps.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235296.4657884-1016-84239282964901/.source.nft follow=False _original_basename=jump-chain.j2 checksum=4c6f036d2d5808f109acc0880c19aa74ca48c961 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9046 DF PROTO=TCP SPT=49984 DPT=9101 SEQ=262484235 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288A1DA0000000001030307) Nov 27 04:21:38 localhost python3.9[143369]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-update-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46714 DF PROTO=TCP SPT=58200 DPT=9882 SEQ=3329947740 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288A4960000000001030307) Nov 27 04:21:38 localhost python3.9[143442]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-update-jumps.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235297.7783298-1061-119270285923917/.source.nft follow=False _original_basename=jump-chain.j2 checksum=4c6f036d2d5808f109acc0880c19aa74ca48c961 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:39 localhost python3.9[143534]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-flushes.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:40 localhost python3.9[143607]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-flushes.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235299.0639305-1107-254247871771871/.source.nft follow=False _original_basename=flush-chain.j2 checksum=d16337256a56373421842284fe09e4e6c7df417e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:40 localhost python3.9[143699]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-chains.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:41 localhost python3.9[143772]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-chains.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235300.3562238-1151-131009841799552/.source.nft follow=False _original_basename=chains.j2 checksum=2079f3b60590a165d1d502e763170876fc8e2984 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:42 localhost python3.9[143864]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-rules.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:21:42 localhost python3.9[143937]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-rules.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235301.5501633-1196-96624395877099/.source.nft follow=False _original_basename=ruleset.j2 checksum=15a82a0dc61abfd6aa593407582b5b950437eb80 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:42 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57748 DF PROTO=TCP SPT=33270 DPT=9882 SEQ=3887614405 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288B3DB0000000001030307) Nov 27 04:21:43 localhost python3.9[144029]: ansible-ansible.builtin.file Invoked with group=root mode=0600 owner=root path=/etc/nftables/edpm-rules.nft.changed state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:43 localhost python3.9[144121]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-chains.nft /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft /etc/nftables/edpm-jumps.nft | nft -c -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:21:45 localhost python3.9[144216]: ansible-ansible.builtin.blockinfile Invoked with backup=False block=include "/etc/nftables/iptables.nft"#012include "/etc/nftables/edpm-chains.nft"#012include "/etc/nftables/edpm-rules.nft"#012include "/etc/nftables/edpm-jumps.nft"#012 path=/etc/sysconfig/nftables.conf validate=nft -c -f %s state=present marker=# {mark} ANSIBLE MANAGED BLOCK create=False marker_begin=BEGIN marker_end=END append_newline=False prepend_newline=False encoding=utf-8 unsafe_writes=False insertafter=None insertbefore=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42764 DF PROTO=TCP SPT=53336 DPT=9102 SEQ=4271511787 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288BF2A0000000001030307) Nov 27 04:21:46 localhost python3.9[144309]: ansible-ansible.builtin.file Invoked with group=hugetlbfs mode=0775 owner=zuul path=/dev/hugepages1G state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:47 localhost python3.9[144401]: ansible-ansible.builtin.file Invoked with group=hugetlbfs mode=0775 owner=zuul path=/dev/hugepages2M state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:21:48 localhost python3.9[144493]: ansible-ansible.posix.mount Invoked with fstype=hugetlbfs opts=pagesize=1G path=/dev/hugepages1G src=none state=mounted boot=True dump=0 opts_no_log=False passno=0 backup=False fstab=None Nov 27 04:21:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62398 DF PROTO=TCP SPT=49600 DPT=9100 SEQ=3728545456 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288CBDB0000000001030307) Nov 27 04:21:49 localhost python3.9[144586]: ansible-ansible.posix.mount Invoked with fstype=hugetlbfs opts=pagesize=2M path=/dev/hugepages2M src=none state=mounted boot=True dump=0 opts_no_log=False passno=0 backup=False fstab=None Nov 27 04:21:49 localhost systemd-logind[761]: Session 44 logged out. Waiting for processes to exit. Nov 27 04:21:49 localhost systemd[1]: session-44.scope: Deactivated successfully. Nov 27 04:21:49 localhost systemd[1]: session-44.scope: Consumed 28.288s CPU time. Nov 27 04:21:49 localhost systemd-logind[761]: Removed session 44. Nov 27 04:21:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55673 DF PROTO=TCP SPT=46326 DPT=9101 SEQ=2164377654 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288DAF30000000001030307) Nov 27 04:21:55 localhost sshd[144602]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:21:55 localhost systemd-logind[761]: New session 45 of user zuul. Nov 27 04:21:55 localhost systemd[1]: Started Session 45 of User zuul. Nov 27 04:21:55 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9778 DF PROTO=TCP SPT=47944 DPT=9105 SEQ=1901118972 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288E7150000000001030307) Nov 27 04:21:56 localhost python3.9[144697]: ansible-ansible.builtin.tempfile Invoked with state=file prefix=ansible. suffix= path=None Nov 27 04:21:57 localhost systemd[1]: systemd-timedated.service: Deactivated successfully. Nov 27 04:21:57 localhost python3.9[144792]: ansible-ansible.builtin.stat Invoked with path=/etc/ssh/ssh_known_hosts follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:21:59 localhost python3.9[144948]: ansible-ansible.builtin.slurp Invoked with src=/etc/ssh/ssh_known_hosts Nov 27 04:22:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41846 DF PROTO=TCP SPT=35360 DPT=9105 SEQ=1657358675 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1288F7DB0000000001030307) Nov 27 04:22:00 localhost python3.9[145055]: ansible-ansible.legacy.stat Invoked with path=/tmp/ansible.sqdb75iy follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:22:01 localhost python3.9[145130]: ansible-ansible.legacy.copy Invoked with dest=/tmp/ansible.sqdb75iy mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235320.3304558-191-273544906680557/.source.sqdb75iy _original_basename=.dw0hukz3 follow=False checksum=e4747bb80fc7defd6e17ae91494a027cfd629844 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:03 localhost python3.9[145222]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'ssh_host_key_rsa_public', 'ssh_host_key_ed25519_public', 'ssh_host_key_ecdsa_public'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:22:05 localhost python3.9[145314]: ansible-ansible.builtin.blockinfile Invoked with block=np0005537442.localdomain,192.168.122.104,np0005537442* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDZ6+t2RQvYi/f60sV6qGAF60A1plHYLpQn7dD2IhG+Czr5OEjxITemq/7OJS7h+Ne9DU0X2SWSz0GMHkvJzD8MWBUTqEiZSsKGpTU5TzJgBbtQh4a/LJPLTMVeWh1bGNqmhfMab+H676ZaRem42hcIaqksBkctHcH+vYEugPUULNxUUYotCsiDPtQcGY4sB0PAH+TXLOLsdSLlWFAn8SaZLGixyiXxQzRJBBewu5KDE3jRFntZPNjAZUmgOKUNHTkNCwTCURrVkjOqTx2h4cbJFLTgH5cua9N1SClN7wCHHd9nOZKZJokrHltXDSxGfdSMM8LnQ0vl8/tkQ0mZxST1LVzUcUeTW+kNz44ygmPbBpR3F4sWj4zM+miT/2EarGnH87pJ3NeiNNp2ZEBx9uoowFcdECFaIBSf8+29k3AdB5tFYsGzkKQ81cZjNsC8bjDU3LNufSmtl922xHJuXclY1wsdLHJjerf5NNBh4V/5DzzBOC8tQuuJcIlUBd8mfX0=#012np0005537442.localdomain,192.168.122.104,np0005537442* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGI0WfjNJuUldmVezz3VNZsRFBhLxd7PtEMnK4wnAkZy#012np0005537442.localdomain,192.168.122.104,np0005537442* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHe4xzDMYg0Ds+ttH0MQh9kxHG1FIw1Ow0NzwB4ZYElnIONX6tamwfiH2ucNJX2T+QASDUPQlJUm/eRd8Lg5q6A=#012np0005537444.localdomain,192.168.122.106,np0005537444* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVQnNmX1+68w+ZbvmHMSJP3G6Gp5eoU/UG1TPEqifZH8HzinCXMNkeyCf3BVotH4CvjVs1UyzcJAGbOj+4N9NMuosvJcf7WTfQSxAkJB9Dz3gV9yg/LAScQsvVi+t9BxPkiPLrjs/JSyy4K6D0LQ8TLlXrhcrwVTHQvGKPvJqlD2RO+kLhWz/jR7en1FPLamRhvW3w6dAWI+nYsKUUGDdu6lsb2nHwAEyIHnLp+IeCbeM/apPU9DSh6ERNERW8QAqmBIDeEp3C1sNEtiBMVtj9fAH7SqTI/fRnJTD1LP6ziSGlVXgFJdkCs4lPmNuOEagV9DDO7l4eD0kzAJ1o/a8uvUDkRBHjrRExUxVeaOH2hNyNY5sZ0Cw9CZXScubJJkxlKbMJuzHw3ZpvoECi1yaJVCYiQ/1JznRtXbEHwBstIiCLmrUtEiMoZ7oH45v8B9yHNHdtuxZQHYybx+4JFrv2uqtUEhJ2QmWGqTIQGAA0ITLL6NznaorLZDpDkMuDBXs=#012np0005537444.localdomain,192.168.122.106,np0005537444* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILHjN1rwuss+VqaF46hoZaGgDybemAkesm9IW+B3R/U0#012np0005537444.localdomain,192.168.122.106,np0005537444* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLOU2xXQKCM6J7Aff6uB5gbI8C9ZImkfOj3Vt4FSDq62HRqP5tp02aUzOhc1LIfbyaQY3pzABROZsz3vKsP4i/g=#012np0005537446.localdomain,192.168.122.108,np0005537446* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZeWh6gZHJpy+jtMRG/XGdhy9SYywGs3q0MIBAiWgx7/hzkaCQxjMMvFVu6G3hROUk6sKQGhF8WPWsE+0AjjsrUl2xCmaViqBT1R0jfZC2U19+mKEhJVUwet4N5BMXRm5YziIc6/BNWGrQ+7gPl1K94zxP8dR4FoSEhlGn2HV4p1iKBt5Z93ZjsRERdN4vUP78W0UfsxStluowzwz5Nu++6mg/D64q1n+arRV9mW3/rHAU1FHGhhK820Y7T3PQgxdBx9GjP7EwDoWITYATGCfWS1a/fMKs5faBezOWJ5dEo7IOUZILjEuOvEJQMCgEsPwUd8G3q0T0uq1BMvdjp8iRhP7hp4ggEB/ciIP3zgnw0g/fsbvQ62pqKs7yWr1jt17deQInxZK+U6oq3Np67FCX8kfJxaAOB98JarjtuJVhrAWhPCRvldwkuqJ7QliNlo4h9ceQAvZNoYMSPOVDh1UwstPXUxyp0l5AQBGqU8jzi3Mt9S0gFTkG1FsGcwWGL6U=#012np0005537446.localdomain,192.168.122.108,np0005537446* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGLMPSi4cyQYzYDgFo1IkVP5PyF7gksrHTPuWudCr2ht#012np0005537446.localdomain,192.168.122.108,np0005537446* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEdY4bB2u9la3fnIsjEGjc2agMqUKyt1+aaOozbPDjQjn7nV76bG5DrOxKJTv4UgxWKVzyf3/etIFL5wlVgIshc=#012np0005537445.localdomain,192.168.122.107,np0005537445* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1/piBT2PI0lEnlBA9RLkYttkVms78XEr2pK4jrGzXLJsjY5YhAhpZjABbX6IrEXb0jtLr9fu2OK3NtHKlrjNb/7TkxYw4YuiZWvdYFf4s6Q4FTcCO6oLrFLcEgu5Rzzmv0bCZSCXSH7ZrDLbW3KwkiwaTCUIu9j3GxfcmFYjiTzDPQTsskQc06ojYDvAc0mflNdibjhZHmD73Yu4NM5EtulvOe7cIAF5fk6E8y41QtFlDZVKun2C2copAdORw4q1KMoxBeHLUDgGQniBbhgct43//SHZpvNiSjsvvV1CyEotNjbAd1nyuAiHMDy9mON8NSR4hJfrBRXNDNmenoJNQErYdcIJQZwiXMZEGzVgK2LPcTnlyGKMwZdailMhaZ7vqxYN9TxRRUxsopDhhk4ekYF9FtxBbt2KfRzjqQOkmiiNaYAiBdUCd/ogfYfRl54F0j/FGmrljjHOw0oPcQVd/cZiwZrgjwnyoMahxDpiCRE21ptokmRC8f24fhQXEJ80=#012np0005537445.localdomain,192.168.122.107,np0005537445* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDW1YU1OdWijXeHVqrnhAcz+NyJCNpix91kORj9tKwLi#012np0005537445.localdomain,192.168.122.107,np0005537445* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBIGflEqBl99hsm/7PM0XMC1qp27xIy8pRpS+301RlLLxmLwsXVWsTn7OOAEc9LH/NrCDCf4HPS2+dx+DYmhsm8=#012np0005537441.localdomain,192.168.122.103,np0005537441* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDWRAei0LGbei1vF4gcoA5zoYcENUUXbT+RNxRrzeTBr84EZ6DQ6p9T5XNKUcEoe7tPEv7QsuZekWJ50wtUWSZRlQCeG+B/cTNlvRx3tfOciPFmCM7x149Oo3xTUEk+TxUgg+U31VxWmmVLod2YnhscV+n5AOSoCWF6F8PqLVJI1q7TQYzmKK5C4QkD7eNwDBkMDHaeWMnXOdkw025dKxRCzE7BttEEmXznEskx6l9MQR432QsxjGBhUNCvHYwG6/fm31ELhOX1BITpI5DzQu3bjnxHLTSgpmvP5Y6AgXn/Snh0394S4PeMDEzK2ivAMuC0mG/dPet7nDKlY8QZrsd8TH4a9REwvZ3DKXbYx5odMeB357AyW1CsKniltq/Icq4sOES9STUZEowOlFZfcqc9Oh7o89+vEvORNNKcLMq91qDTlBFBFDjsPYvkCzhCSlxUCZ1AICp6prURQDrfmwM881TUKqMZsgnOmXjz/TWlE5e/XW5fmXrRTxMFljrpYgk=#012np0005537441.localdomain,192.168.122.103,np0005537441* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJyJ5p0fTUeXalGTImEC7YqoR45jkTfcdtSe3cJbPdoB#012np0005537441.localdomain,192.168.122.103,np0005537441* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEGrHJlz0p658nB+Qc1eEFqby09RjLzKtmG0hspFBzR8xqpTjyemadpExyOhEdJX8w7Ot/13IkavrZpAp3IfiGw=#012np0005537443.localdomain,192.168.122.105,np0005537443* ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+TzqlkhU2yG8BpCyg9Jf4bhGZbiRnL+Jc1FYzYFLOntd5BBvaw6hhYNzDc/mKZeP0tlyeFmf0aA+wYpdWrTwIgTPqGx4UhnipnEPZSEFfJn8o0fzHqgZ5h22y7XkklotRytWOPX87vHDdeYDciC1tnzCEdxTJMYwBzM41MaEXcO+Y517qGhLL8ftIsAjgvw9GK5z6h1iEA66zTONVmi/I6HZhjaRkhUmhi4oQ39ipF6wSAd79aQikmsuUq2NBHKGbvonjGPTI6ni5CmjVWa4dTpua5+pCHgJAjYmrD+nZvAPxHPdz2Z4H9rlYsmC2An0sOXROsGuHEeQzoSPmVbXTH9m865cib1g3r+C6usVsK4pVA6pS8TEXkAEJ61rQWYLdxykz52kixRP242TnVJU+znQgKsgJWcD7997B6BNLZwgB2zSXz+uJRVgQvUQ1ktXL/TtVO+OUrnrBEUzIxZju/ukUoFQyWR8IXwX6U9d0ozh7cXKp/bZ2ah9U0EWnLTc=#012np0005537443.localdomain,192.168.122.105,np0005537443* ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOrf1oxWc5mRHVe0pMaoKKhhYAO3G19WZtnxShtTpvT7#012np0005537443.localdomain,192.168.122.105,np0005537443* ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBP2FXCzWlsUIEx8iA14IH5Igdj3o29TGgHNXSBWi1wqBOd0IbhwWuDQlu3SaqJth++sAktcVUO+V/x08aleyOhc=#012 create=True mode=0644 path=/tmp/ansible.sqdb75iy state=present marker=# {mark} ANSIBLE MANAGED BLOCK backup=False marker_begin=BEGIN marker_end=END append_newline=False prepend_newline=False encoding=utf-8 unsafe_writes=False insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:06 localhost python3.9[145406]: ansible-ansible.legacy.command Invoked with _raw_params=cat '/tmp/ansible.sqdb75iy' > /etc/ssh/ssh_known_hosts _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:22:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37141 DF PROTO=TCP SPT=51498 DPT=9882 SEQ=1884473795 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128919C70000000001030307) Nov 27 04:22:08 localhost python3.9[145500]: ansible-ansible.builtin.file Invoked with path=/tmp/ansible.sqdb75iy state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:09 localhost systemd[1]: session-45.scope: Deactivated successfully. Nov 27 04:22:09 localhost systemd[1]: session-45.scope: Consumed 4.178s CPU time. Nov 27 04:22:09 localhost systemd-logind[761]: Session 45 logged out. Waiting for processes to exit. Nov 27 04:22:09 localhost systemd-logind[761]: Removed session 45. Nov 27 04:22:14 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39767 DF PROTO=TCP SPT=34442 DPT=9100 SEQ=3421039947 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128930810000000001030307) Nov 27 04:22:15 localhost sshd[145515]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:22:15 localhost systemd-logind[761]: New session 46 of user zuul. Nov 27 04:22:15 localhost systemd[1]: Started Session 46 of User zuul. Nov 27 04:22:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22718 DF PROTO=TCP SPT=39464 DPT=9102 SEQ=1468449363 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128934590000000001030307) Nov 27 04:22:16 localhost python3.9[145608]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:22:18 localhost python3.9[145704]: ansible-ansible.builtin.systemd Invoked with enabled=True name=sshd daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 04:22:19 localhost python3.9[145798]: ansible-ansible.builtin.systemd Invoked with name=sshd state=started daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:22:20 localhost python3.9[145891]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/edpm-chains.nft _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:22:21 localhost python3.9[145984]: ansible-ansible.builtin.stat Invoked with path=/etc/nftables/edpm-rules.nft.changed follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:22:22 localhost python3.9[146078]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft | nft -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:22:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2891 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128950240000000001030307) Nov 27 04:22:23 localhost python3.9[146173]: ansible-ansible.builtin.file Invoked with path=/etc/nftables/edpm-rules.nft.changed state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2892 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289541A0000000001030307) Nov 27 04:22:23 localhost systemd[1]: session-46.scope: Deactivated successfully. Nov 27 04:22:23 localhost systemd[1]: session-46.scope: Consumed 3.962s CPU time. Nov 27 04:22:23 localhost systemd-logind[761]: Session 46 logged out. Waiting for processes to exit. Nov 27 04:22:23 localhost systemd-logind[761]: Removed session 46. Nov 27 04:22:25 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2893 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12895C1A0000000001030307) Nov 27 04:22:25 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2067 DF PROTO=TCP SPT=39390 DPT=9105 SEQ=19547553 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12895C440000000001030307) Nov 27 04:22:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2068 DF PROTO=TCP SPT=39390 DPT=9105 SEQ=19547553 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289605A0000000001030307) Nov 27 04:22:28 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2069 DF PROTO=TCP SPT=39390 DPT=9105 SEQ=19547553 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289685A0000000001030307) Nov 27 04:22:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2894 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12896BDA0000000001030307) Nov 27 04:22:30 localhost sshd[146188]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:22:30 localhost systemd-logind[761]: New session 47 of user zuul. Nov 27 04:22:30 localhost systemd[1]: Started Session 47 of User zuul. Nov 27 04:22:31 localhost python3.9[146281]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:22:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2070 DF PROTO=TCP SPT=39390 DPT=9105 SEQ=19547553 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289781B0000000001030307) Nov 27 04:22:33 localhost python3.9[146377]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:22:34 localhost python3.9[146431]: ansible-ansible.legacy.dnf Invoked with name=['yum-utils'] allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None state=None Nov 27 04:22:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2895 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12898BDB0000000001030307) Nov 27 04:22:38 localhost python3.9[146523]: ansible-ansible.legacy.command Invoked with _raw_params=needs-restarting -r _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:22:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19142 DF PROTO=TCP SPT=51596 DPT=9882 SEQ=8808228 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12898EF70000000001030307) Nov 27 04:22:39 localhost python3.9[146616]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/openstack/reboot_required/ state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:39 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19143 DF PROTO=TCP SPT=51596 DPT=9882 SEQ=8808228 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289931B0000000001030307) Nov 27 04:22:40 localhost python3.9[146708]: ansible-ansible.builtin.file Invoked with mode=0600 path=/var/lib/openstack/reboot_required/needs_restarting state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:41 localhost python3.9[146800]: ansible-ansible.builtin.lineinfile Invoked with dest=/var/lib/openstack/reboot_required/needs_restarting line=Not root, Subscription Management repositories not updated#012Core libraries or services have been updated since boot-up:#012 * systemd#012#012Reboot is required to fully utilize these updates.#012More information: https://access.redhat.com/solutions/27943 path=/var/lib/openstack/reboot_required/needs_restarting state=present encoding=utf-8 backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19144 DF PROTO=TCP SPT=51596 DPT=9882 SEQ=8808228 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12899B1A0000000001030307) Nov 27 04:22:41 localhost python3.9[146890]: ansible-ansible.builtin.find Invoked with paths=['/var/lib/openstack/reboot_required/'] patterns=[] read_whole_file=False file_type=file age_stamp=mtime recurse=False hidden=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:22:43 localhost python3.9[146980]: ansible-ansible.builtin.stat Invoked with path=/var/lib/config-data/puppet-generated follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:22:44 localhost python3.9[147072]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/config follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:22:44 localhost systemd[1]: session-47.scope: Deactivated successfully. Nov 27 04:22:44 localhost systemd[1]: session-47.scope: Consumed 8.888s CPU time. Nov 27 04:22:44 localhost systemd-logind[761]: Session 47 logged out. Waiting for processes to exit. Nov 27 04:22:44 localhost systemd-logind[761]: Removed session 47. Nov 27 04:22:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42594 DF PROTO=TCP SPT=44814 DPT=9102 SEQ=1619728100 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289A9890000000001030307) Nov 27 04:22:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20488 DF PROTO=TCP SPT=50924 DPT=9100 SEQ=658167320 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289B19A0000000001030307) Nov 27 04:22:50 localhost sshd[147087]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:22:50 localhost systemd-logind[761]: New session 48 of user zuul. Nov 27 04:22:50 localhost systemd[1]: Started Session 48 of User zuul. Nov 27 04:22:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20489 DF PROTO=TCP SPT=50924 DPT=9100 SEQ=658167320 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289C15B0000000001030307) Nov 27 04:22:51 localhost python3.9[147180]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:22:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11614 DF PROTO=TCP SPT=38098 DPT=9101 SEQ=402776948 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289C95A0000000001030307) Nov 27 04:22:54 localhost python3.9[147276]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/telemetry setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:22:54 localhost python3.9[147368]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:22:55 localhost python3.9[147441]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235374.2757912-183-24589578489968/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:56 localhost python3.9[147533]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/neutron-sriov setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:22:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19286 DF PROTO=TCP SPT=58434 DPT=9105 SEQ=910382256 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289D59A0000000001030307) Nov 27 04:22:56 localhost python3.9[147625]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/neutron-sriov/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:22:57 localhost python3.9[147698]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/neutron-sriov/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235376.3989656-253-250660108411994/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:58 localhost python3.9[147790]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/neutron-dhcp setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:22:58 localhost python3.9[147882]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/neutron-dhcp/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:22:59 localhost python3.9[147955]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/neutron-dhcp/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235378.2553658-326-87233315570056/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:22:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11616 DF PROTO=TCP SPT=38098 DPT=9101 SEQ=402776948 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289E11B0000000001030307) Nov 27 04:23:00 localhost chronyd[139282]: Selected source 54.39.23.64 (pool.ntp.org) Nov 27 04:23:00 localhost python3.9[148089]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:00 localhost systemd[1]: tmp-crun.ZPIkFb.mount: Deactivated successfully. Nov 27 04:23:00 localhost podman[148193]: 2025-11-27 09:23:00.973619572 +0000 UTC m=+0.098122415 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, distribution-scope=public, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , version=7, description=Red Hat Ceph Storage 7, vcs-type=git, name=rhceph, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:23:01 localhost podman[148193]: 2025-11-27 09:23:01.065265291 +0000 UTC m=+0.189768134 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, architecture=x86_64, version=7, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, io.openshift.expose-services=, GIT_CLEAN=True, distribution-scope=public, io.buildah.version=1.33.12, ceph=True, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, GIT_BRANCH=main) Nov 27 04:23:01 localhost python3.9[148278]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:01 localhost python3.9[148413]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235380.8171363-396-97151704575387/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=19288 DF PROTO=TCP SPT=58434 DPT=9105 SEQ=910382256 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1289ED5A0000000001030307) Nov 27 04:23:02 localhost python3.9[148536]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:03 localhost python3.9[148643]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/libvirt/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:04 localhost python3.9[148716]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/libvirt/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235383.1047025-468-192594275810710/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:04 localhost python3.9[148808]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/ovn setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:05 localhost python3.9[148900]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:05 localhost python3.9[148973]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235384.9312766-541-93308730087039/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:06 localhost python3.9[149065]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/bootstrap setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:07 localhost python3.9[149157]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/bootstrap/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:07 localhost python3.9[149230]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/bootstrap/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235386.7860136-613-27144408536168/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11617 DF PROTO=TCP SPT=38098 DPT=9101 SEQ=402776948 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A01DA0000000001030307) Nov 27 04:23:08 localhost python3.9[149322]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/var/lib/openstack/cacerts/neutron-metadata setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31039 DF PROTO=TCP SPT=51966 DPT=9882 SEQ=65780620 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A04270000000001030307) Nov 27 04:23:09 localhost python3.9[149414]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:09 localhost python3.9[149487]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235388.616836-685-3243886323415/.source.pem _original_basename=tls-ca-bundle.pem follow=False checksum=37c56a6007cd7a43a1e4fca58da077efc3fcf708 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:10 localhost systemd[1]: session-48.scope: Deactivated successfully. Nov 27 04:23:10 localhost systemd[1]: session-48.scope: Consumed 11.752s CPU time. Nov 27 04:23:10 localhost systemd-logind[761]: Session 48 logged out. Waiting for processes to exit. Nov 27 04:23:10 localhost systemd-logind[761]: Removed session 48. Nov 27 04:23:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31041 DF PROTO=TCP SPT=51966 DPT=9882 SEQ=65780620 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A101B0000000001030307) Nov 27 04:23:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24245 DF PROTO=TCP SPT=60112 DPT=9102 SEQ=3306810121 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A1EB90000000001030307) Nov 27 04:23:15 localhost sshd[149502]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:23:16 localhost systemd-logind[761]: New session 49 of user zuul. Nov 27 04:23:16 localhost systemd[1]: Started Session 49 of User zuul. Nov 27 04:23:16 localhost python3.9[149597]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/openstack/config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9579 DF PROTO=TCP SPT=46696 DPT=9100 SEQ=2813053571 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A26DB0000000001030307) Nov 27 04:23:17 localhost python3.9[149689]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/ceph/ceph.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:18 localhost python3.9[149762]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/ceph/ceph.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235397.1378937-64-53147606326962/.source.conf _original_basename=ceph.conf follow=False checksum=059f91412ca8499ae565c4f9f7060e76ecdacc99 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:19 localhost python3.9[149854]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/ceph/ceph.client.openstack.keyring follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:19 localhost python3.9[149927]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/ceph/ceph.client.openstack.keyring mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235398.5727642-64-235936512573038/.source.keyring _original_basename=ceph.client.openstack.keyring follow=False checksum=4387bc148c9a58f9a0b1eb236093f27263bcb30f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:20 localhost systemd-logind[761]: Session 49 logged out. Waiting for processes to exit. Nov 27 04:23:20 localhost systemd[1]: session-49.scope: Deactivated successfully. Nov 27 04:23:20 localhost systemd[1]: session-49.scope: Consumed 2.322s CPU time. Nov 27 04:23:20 localhost systemd-logind[761]: Removed session 49. Nov 27 04:23:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9580 DF PROTO=TCP SPT=46696 DPT=9100 SEQ=2813053571 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A369A0000000001030307) Nov 27 04:23:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24549 DF PROTO=TCP SPT=58012 DPT=9101 SEQ=2199303475 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A3E9A0000000001030307) Nov 27 04:23:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2897 DF PROTO=TCP SPT=51574 DPT=9101 SEQ=784295511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A49DB0000000001030307) Nov 27 04:23:27 localhost sshd[149942]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:23:27 localhost systemd-logind[761]: New session 50 of user zuul. Nov 27 04:23:27 localhost systemd[1]: Started Session 50 of User zuul. Nov 27 04:23:28 localhost python3.9[150035]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:23:29 localhost python3.9[150131]: ansible-ansible.builtin.file Invoked with group=zuul mode=0750 owner=zuul path=/var/lib/edpm-config/firewall setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2073 DF PROTO=TCP SPT=39390 DPT=9105 SEQ=19547553 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A55DA0000000001030307) Nov 27 04:23:30 localhost python3.9[150223]: ansible-ansible.builtin.file Invoked with group=openvswitch owner=openvswitch path=/var/lib/openvswitch/ovn setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:23:30 localhost python3.9[150313]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'selinux'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:23:31 localhost python3.9[150405]: ansible-ansible.posix.seboolean Invoked with name=virt_sandbox_use_netlink persistent=True state=True ignore_selinux_state=False Nov 27 04:23:32 localhost python3.9[150497]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:23:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15987 DF PROTO=TCP SPT=48860 DPT=9105 SEQ=2910675343 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A625B0000000001030307) Nov 27 04:23:33 localhost python3.9[150551]: ansible-ansible.legacy.dnf Invoked with name=['openvswitch3.3'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:23:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24552 DF PROTO=TCP SPT=58012 DPT=9101 SEQ=2199303475 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A75DA0000000001030307) Nov 27 04:23:38 localhost python3.9[150645]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=openvswitch.service state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:23:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31429 DF PROTO=TCP SPT=54500 DPT=9882 SEQ=1435434905 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A79570000000001030307) Nov 27 04:23:40 localhost python3[150740]: ansible-osp.edpm.edpm_nftables_snippet Invoked with content=- rule_name: 118 neutron vxlan networks#012 rule:#012 proto: udp#012 dport: 4789#012- rule_name: 119 neutron geneve networks#012 rule:#012 proto: udp#012 dport: 6081#012 state: ["UNTRACKED"]#012- rule_name: 120 neutron geneve networks no conntrack#012 rule:#012 proto: udp#012 dport: 6081#012 table: raw#012 chain: OUTPUT#012 jump: NOTRACK#012 action: append#012 state: []#012- rule_name: 121 neutron geneve networks no conntrack#012 rule:#012 proto: udp#012 dport: 6081#012 table: raw#012 chain: PREROUTING#012 jump: NOTRACK#012 action: append#012 state: []#012 dest=/var/lib/edpm-config/firewall/ovn.yaml state=present Nov 27 04:23:40 localhost python3.9[150832]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/var/lib/edpm-config/firewall state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:41 localhost python3.9[150924]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31431 DF PROTO=TCP SPT=54500 DPT=9882 SEQ=1435434905 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A855A0000000001030307) Nov 27 04:23:42 localhost python3.9[150972]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml _original_basename=base-rules.yaml.j2 recurse=False state=file path=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:42 localhost python3.9[151064]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:43 localhost python3.9[151112]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml _original_basename=.dnqbgcey recurse=False state=file path=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:43 localhost python3.9[151204]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/iptables.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:44 localhost python3.9[151252]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/iptables.nft _original_basename=iptables.nft recurse=False state=file path=/etc/nftables/iptables.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:45 localhost python3.9[151344]: ansible-ansible.legacy.command Invoked with _raw_params=nft -j list ruleset _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:23:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25890 DF PROTO=TCP SPT=33780 DPT=9102 SEQ=2263025883 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A93E90000000001030307) Nov 27 04:23:45 localhost python3[151437]: ansible-edpm_nftables_from_files Invoked with src=/var/lib/edpm-config/firewall Nov 27 04:23:47 localhost python3.9[151529]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24250 DF PROTO=TCP SPT=60112 DPT=9102 SEQ=3306810121 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128A9BDA0000000001030307) Nov 27 04:23:47 localhost python3.9[151604]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-jumps.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235426.1385844-434-272430169351131/.source.nft follow=False _original_basename=jump-chain.j2 checksum=81c2fc96c23335ffe374f9b064e885d5d971ddf9 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:48 localhost python3.9[151696]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-update-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:49 localhost python3.9[151771]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-update-jumps.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235428.1286073-478-9918535367971/.source.nft follow=False _original_basename=jump-chain.j2 checksum=81c2fc96c23335ffe374f9b064e885d5d971ddf9 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:50 localhost python3.9[151863]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-flushes.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:51 localhost python3.9[151938]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-flushes.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235429.984578-523-280697597813445/.source.nft follow=False _original_basename=flush-chain.j2 checksum=4d3ffec49c8eb1a9b80d2f1e8cd64070063a87b4 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=65148 DF PROTO=TCP SPT=45564 DPT=9100 SEQ=1041058596 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AABDA0000000001030307) Nov 27 04:23:51 localhost python3.9[152030]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-chains.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:52 localhost python3.9[152105]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-chains.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235431.2942142-569-197167147312631/.source.nft follow=False _original_basename=chains.j2 checksum=298ada419730ec15df17ded0cc50c97a4014a591 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:53 localhost python3.9[152197]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-rules.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:23:53 localhost python3.9[152272]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-rules.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235432.5780926-614-254245029945564/.source.nft follow=False _original_basename=ruleset.j2 checksum=eb691bdb7d792c5f8ff0d719e807fe1c95b09438 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47066 DF PROTO=TCP SPT=58974 DPT=9101 SEQ=2634355822 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AB3DA0000000001030307) Nov 27 04:23:54 localhost python3.9[152364]: ansible-ansible.builtin.file Invoked with group=root mode=0600 owner=root path=/etc/nftables/edpm-rules.nft.changed state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:54 localhost python3.9[152456]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-chains.nft /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft /etc/nftables/edpm-jumps.nft | nft -c -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:23:55 localhost python3.9[152551]: ansible-ansible.builtin.blockinfile Invoked with backup=False block=include "/etc/nftables/iptables.nft"#012include "/etc/nftables/edpm-chains.nft"#012include "/etc/nftables/edpm-rules.nft"#012include "/etc/nftables/edpm-jumps.nft"#012 path=/etc/sysconfig/nftables.conf validate=nft -c -f %s state=present marker=# {mark} ANSIBLE MANAGED BLOCK create=False marker_begin=BEGIN marker_end=END append_newline=False prepend_newline=False encoding=utf-8 unsafe_writes=False insertafter=None insertbefore=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:56 localhost python3.9[152644]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/edpm-chains.nft _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:23:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63426 DF PROTO=TCP SPT=42360 DPT=9105 SEQ=3193687268 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128ABFDB0000000001030307) Nov 27 04:23:57 localhost python3.9[152737]: ansible-ansible.builtin.stat Invoked with path=/etc/nftables/edpm-rules.nft.changed follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:23:57 localhost python3.9[152831]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft | nft -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:23:58 localhost python3.9[152926]: ansible-ansible.builtin.file Invoked with path=/etc/nftables/edpm-rules.nft.changed state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:23:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47068 DF PROTO=TCP SPT=58974 DPT=9101 SEQ=2634355822 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128ACB9A0000000001030307) Nov 27 04:24:00 localhost python3.9[153016]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'machine'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:24:01 localhost python3.9[153109]: ansible-ansible.legacy.command Invoked with _raw_params=ovs-vsctl set open . external_ids:hostname=np0005537446.localdomain external_ids:ovn-bridge=br-int external_ids:ovn-bridge-mappings=datacentre:br-ex external_ids:ovn-chassis-mac-mappings="datacentre:0e:0a:12:93:5d:7c" external_ids:ovn-encap-ip=172.19.0.108 external_ids:ovn-encap-type=geneve external_ids:ovn-encap-tos=0 external_ids:ovn-match-northd-version=False external_ids:ovn-monitor-all=True external_ids:ovn-remote=tcp:ovsdbserver-sb.openstack.svc:6642 external_ids:ovn-remote-probe-interval=60000 external_ids:ovn-ofctrl-wait-before-clear=8000 external_ids:rundir=/var/run/openvswitch #012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:24:01 localhost ovs-vsctl[153110]: ovs|00001|vsctl|INFO|Called as ovs-vsctl set open . external_ids:hostname=np0005537446.localdomain external_ids:ovn-bridge=br-int external_ids:ovn-bridge-mappings=datacentre:br-ex external_ids:ovn-chassis-mac-mappings=datacentre:0e:0a:12:93:5d:7c external_ids:ovn-encap-ip=172.19.0.108 external_ids:ovn-encap-type=geneve external_ids:ovn-encap-tos=0 external_ids:ovn-match-northd-version=False external_ids:ovn-monitor-all=True external_ids:ovn-remote=tcp:ovsdbserver-sb.openstack.svc:6642 external_ids:ovn-remote-probe-interval=60000 external_ids:ovn-ofctrl-wait-before-clear=8000 external_ids:rundir=/var/run/openvswitch Nov 27 04:24:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63428 DF PROTO=TCP SPT=42360 DPT=9105 SEQ=3193687268 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AD79A0000000001030307) Nov 27 04:24:03 localhost python3.9[153251]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail#012ovs-vsctl show | grep -q "Manager"#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:24:04 localhost python3.9[153356]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:24:05 localhost python3.9[153465]: ansible-ansible.builtin.file Invoked with path=/var/local/libexec recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:05 localhost python3.9[153557]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-container-shutdown follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:06 localhost python3.9[153606]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-container-shutdown _original_basename=edpm-container-shutdown recurse=False state=file path=/var/local/libexec/edpm-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:06 localhost python3.9[153698]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-start-podman-container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:07 localhost python3.9[153746]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-start-podman-container _original_basename=edpm-start-podman-container recurse=False state=file path=/var/local/libexec/edpm-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:07 localhost python3.9[153838]: ansible-ansible.builtin.file Invoked with mode=420 path=/etc/systemd/system-preset state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47069 DF PROTO=TCP SPT=58974 DPT=9101 SEQ=2634355822 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AEBDA0000000001030307) Nov 27 04:24:08 localhost python3.9[153930]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm-container-shutdown.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62002 DF PROTO=TCP SPT=35488 DPT=9882 SEQ=152679040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AEE870000000001030307) Nov 27 04:24:09 localhost python3.9[153978]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/edpm-container-shutdown.service _original_basename=edpm-container-shutdown-service recurse=False state=file path=/etc/systemd/system/edpm-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:09 localhost python3.9[154070]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:10 localhost python3.9[154118]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-edpm-container-shutdown.preset _original_basename=91-edpm-container-shutdown-preset recurse=False state=file path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:11 localhost python3.9[154210]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm-container-shutdown state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:24:11 localhost systemd[1]: Reloading. Nov 27 04:24:11 localhost systemd-rc-local-generator[154236]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:24:11 localhost systemd-sysv-generator[154240]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:24:11 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:24:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62004 DF PROTO=TCP SPT=35488 DPT=9882 SEQ=152679040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128AFA9A0000000001030307) Nov 27 04:24:12 localhost python3.9[154339]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/netns-placeholder.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:12 localhost python3.9[154387]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/etc/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:13 localhost python3.9[154479]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:14 localhost python3.9[154527]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/etc/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:15 localhost python3.9[154619]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=netns-placeholder state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:24:15 localhost systemd[1]: Reloading. Nov 27 04:24:15 localhost systemd-rc-local-generator[154644]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:24:15 localhost systemd-sysv-generator[154647]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:24:15 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:24:15 localhost systemd[1]: Starting Create netns directory... Nov 27 04:24:15 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 04:24:15 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 04:24:15 localhost systemd[1]: Finished Create netns directory. Nov 27 04:24:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33005 DF PROTO=TCP SPT=47242 DPT=9102 SEQ=1838079232 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B09190000000001030307) Nov 27 04:24:16 localhost python3.9[154755]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/healthchecks setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:16 localhost python3.9[154847]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/ovn_controller/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:17 localhost python3.9[154920]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/ovn_controller/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235456.4638665-1345-209510388482560/.source _original_basename=healthcheck follow=False checksum=4098dd010265fabdf5c26b97d169fc4e575ff457 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33661 DF PROTO=TCP SPT=37702 DPT=9100 SEQ=2869195556 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B115A0000000001030307) Nov 27 04:24:18 localhost python3.9[155012]: ansible-ansible.builtin.file Invoked with path=/var/lib/kolla/config_files recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:19 localhost python3.9[155104]: ansible-ansible.legacy.stat Invoked with path=/var/lib/kolla/config_files/ovn_controller.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:19 localhost python3.9[155179]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/kolla/config_files/ovn_controller.json mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235458.6166372-1420-37206499332545/.source.json _original_basename=.cfoe15g0 follow=False checksum=38f75f59f5c2ef6b5da12297bfd31cd1e97012ac backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:20 localhost python3.9[155271]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/edpm-config/container-startup-config/ovn_controller state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33662 DF PROTO=TCP SPT=37702 DPT=9100 SEQ=2869195556 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B211A0000000001030307) Nov 27 04:24:23 localhost python3.9[155528]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/edpm-config/container-startup-config/ovn_controller config_pattern=*.json debug=False Nov 27 04:24:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46125 DF PROTO=TCP SPT=59040 DPT=9101 SEQ=3176879867 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B28DB0000000001030307) Nov 27 04:24:24 localhost python3.9[155620]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:24:25 localhost python3.9[155712]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Nov 27 04:24:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16165 DF PROTO=TCP SPT=45562 DPT=9105 SEQ=4153479652 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B351A0000000001030307) Nov 27 04:24:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46127 DF PROTO=TCP SPT=59040 DPT=9101 SEQ=3176879867 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B409A0000000001030307) Nov 27 04:24:29 localhost python3[155831]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/edpm-config/container-startup-config/ovn_controller config_id=ovn_controller config_overrides={} config_patterns=*.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:24:30 localhost python3[155831]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "52cb1910f3f090372807028d1c2aea98d2557b1086636469529f290368ecdf69",#012 "Digest": "sha256:7ab0ee81fdc9b162df9b50eb2e264c777d08f90975a442620ec940edabe300b2",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:7ab0ee81fdc9b162df9b50eb2e264c777d08f90975a442620ec940edabe300b2"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:43:38.999472418Z",#012 "Config": {#012 "User": "root",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 345745352,#012 "VirtualSize": 345745352,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/d63efe17da859108a09d9b90626ba0c433787abe209cd4ac755f6ba2a5206671/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/d8443c9fdf039c2367e44e0edbe81c941f30f604c3f1eccc2fc81efb5a97a784/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/d8443c9fdf039c2367e44e0edbe81c941f30f604c3f1eccc2fc81efb5a97a784/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:41a433848ac42a81e513766649f77cfa09e37aae045bcbbb33be77f7cf86edc4",#012 "sha256:055d9012b48b3c8064accd40b6372c79c29fedd85061a710ada00677f88b1db9"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "root",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:37.752912815Z",#012 "created_by": "/bin/sh -c dnf install -y ca-certificates dumb-init glibc-langpack-en procps-ng python3 sudo util-l Nov 27 04:24:30 localhost podman[155880]: 2025-11-27 09:24:30.228550519 +0000 UTC m=+0.090322716 container remove aced366b6edd59b0ffcb223897cbbecc9d7e0180af220432f27ea6818483efb2 (image=registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1, name=ovn_controller, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.openshift.expose-services=, vcs-ref=ae875c168a6ec3400acf0a639b71f4bcc4adf272, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'test': '/openstack/healthcheck 6642'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ovn-controller:17.1', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'user': 'root', 'volumes': ['/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/log/containers/openvswitch:/var/log/openvswitch:z', '/var/log/containers/openvswitch:/var/log/ovn:z']}, container_name=ovn_controller, distribution-scope=public, description=Red Hat OpenStack Platform 17.1 ovn-controller, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ovn-controller, url=https://www.redhat.com, vcs-type=git, io.buildah.version=1.41.4, tcib_managed=true, io.k8s.description=Red Hat OpenStack Platform 17.1 ovn-controller, name=rhosp17/openstack-ovn-controller, vendor=Red Hat, Inc., build-date=2025-11-18T23:34:05Z, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=ae875c168a6ec3400acf0a639b71f4bcc4adf272, release=1761123044, version=17.1.12, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, managed_by=tripleo_ansible, com.redhat.component=openstack-ovn-controller-container, config_id=tripleo_step4, summary=Red Hat OpenStack Platform 17.1 ovn-controller, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ovn-controller, konflux.additional-tags=17.1.12 17.1_20251118.1) Nov 27 04:24:30 localhost python3[155831]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman rm --force ovn_controller Nov 27 04:24:30 localhost podman[155894]: Nov 27 04:24:30 localhost podman[155894]: 2025-11-27 09:24:30.337916067 +0000 UTC m=+0.089536697 container create c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_controller) Nov 27 04:24:30 localhost podman[155894]: 2025-11-27 09:24:30.29516148 +0000 UTC m=+0.046782130 image pull quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified Nov 27 04:24:30 localhost python3[155831]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name ovn_controller --conmon-pidfile /run/ovn_controller.pid --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --healthcheck-command /openstack/healthcheck --label config_id=ovn_controller --label container_name=ovn_controller --label managed_by=edpm_ansible --label config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']} --log-driver journald --log-level info --network host --privileged=True --user root --volume /lib/modules:/lib/modules:ro --volume /run:/run --volume /var/lib/openvswitch/ovn:/run/ovn:shared,z --volume /var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z --volume /var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified Nov 27 04:24:31 localhost python3.9[156022]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:24:32 localhost python3.9[156116]: ansible-file Invoked with path=/etc/systemd/system/edpm_ovn_controller.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:32 localhost python3.9[156162]: ansible-stat Invoked with path=/etc/systemd/system/edpm_ovn_controller_healthcheck.timer follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:24:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16167 DF PROTO=TCP SPT=45562 DPT=9105 SEQ=4153479652 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B4CDA0000000001030307) Nov 27 04:24:33 localhost python3.9[156253]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764235472.4683175-1684-274781716201666/source dest=/etc/systemd/system/edpm_ovn_controller.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:24:33 localhost python3.9[156299]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:24:33 localhost systemd[1]: Reloading. Nov 27 04:24:33 localhost systemd-rc-local-generator[156326]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:24:33 localhost systemd-sysv-generator[156329]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:24:33 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:24:34 localhost python3.9[156380]: ansible-systemd Invoked with state=restarted name=edpm_ovn_controller.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:24:34 localhost systemd[1]: Reloading. Nov 27 04:24:34 localhost systemd-rc-local-generator[156410]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:24:34 localhost systemd-sysv-generator[156413]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:24:34 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:24:34 localhost systemd[1]: Starting ovn_controller container... Nov 27 04:24:35 localhost systemd[1]: Started libcrun container. Nov 27 04:24:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5fb0e0a3b19797174f4e78752405d331b5e7ab034cbe7379d6dac1c2cae5ab12/merged/run/ovn supports timestamps until 2038 (0x7fffffff) Nov 27 04:24:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:24:35 localhost podman[156422]: 2025-11-27 09:24:35.05692086 +0000 UTC m=+0.155527955 container init c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:24:35 localhost ovn_controller[156436]: + sudo -E kolla_set_configs Nov 27 04:24:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:24:35 localhost podman[156422]: 2025-11-27 09:24:35.107973229 +0000 UTC m=+0.206580324 container start c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, managed_by=edpm_ansible) Nov 27 04:24:35 localhost edpm-start-podman-container[156422]: ovn_controller Nov 27 04:24:35 localhost systemd[1]: Created slice User Slice of UID 0. Nov 27 04:24:35 localhost systemd[1]: Starting User Runtime Directory /run/user/0... Nov 27 04:24:35 localhost systemd[1]: Finished User Runtime Directory /run/user/0. Nov 27 04:24:35 localhost systemd[1]: Starting User Manager for UID 0... Nov 27 04:24:35 localhost podman[156444]: 2025-11-27 09:24:35.196671187 +0000 UTC m=+0.090003268 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=starting, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller) Nov 27 04:24:35 localhost podman[156444]: 2025-11-27 09:24:35.205187735 +0000 UTC m=+0.098519816 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_controller, config_id=ovn_controller, org.label-schema.build-date=20251125) Nov 27 04:24:35 localhost podman[156444]: unhealthy Nov 27 04:24:35 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:24:35 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Failed with result 'exit-code'. Nov 27 04:24:35 localhost edpm-start-podman-container[156421]: Creating additional drop-in dependency for "ovn_controller" (c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133) Nov 27 04:24:35 localhost systemd[1]: Reloading. Nov 27 04:24:35 localhost systemd[156466]: Queued start job for default target Main User Target. Nov 27 04:24:35 localhost systemd[156466]: Created slice User Application Slice. Nov 27 04:24:35 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.1 (250 of 333 items), suggesting rotation. Nov 27 04:24:35 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:24:35 localhost systemd[156466]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). Nov 27 04:24:35 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:24:35 localhost systemd[156466]: Started Daily Cleanup of User's Temporary Directories. Nov 27 04:24:35 localhost systemd[156466]: Reached target Paths. Nov 27 04:24:35 localhost systemd[156466]: Reached target Timers. Nov 27 04:24:35 localhost systemd[156466]: Starting D-Bus User Message Bus Socket... Nov 27 04:24:35 localhost systemd[156466]: Starting Create User's Volatile Files and Directories... Nov 27 04:24:35 localhost systemd[156466]: Finished Create User's Volatile Files and Directories. Nov 27 04:24:35 localhost systemd[156466]: Listening on D-Bus User Message Bus Socket. Nov 27 04:24:35 localhost systemd[156466]: Reached target Sockets. Nov 27 04:24:35 localhost systemd[156466]: Reached target Basic System. Nov 27 04:24:35 localhost systemd[156466]: Reached target Main User Target. Nov 27 04:24:35 localhost systemd[156466]: Startup finished in 119ms. Nov 27 04:24:35 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:24:35 localhost systemd-rc-local-generator[156525]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:24:35 localhost systemd-sysv-generator[156530]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:24:35 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:24:35 localhost systemd[1]: Started User Manager for UID 0. Nov 27 04:24:35 localhost systemd[1]: Started ovn_controller container. Nov 27 04:24:35 localhost systemd[1]: Started Session c12 of User root. Nov 27 04:24:35 localhost ovn_controller[156436]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:24:35 localhost ovn_controller[156436]: INFO:__main__:Validating config file Nov 27 04:24:35 localhost ovn_controller[156436]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:24:35 localhost ovn_controller[156436]: INFO:__main__:Writing out command to execute Nov 27 04:24:35 localhost systemd[1]: session-c12.scope: Deactivated successfully. Nov 27 04:24:35 localhost ovn_controller[156436]: ++ cat /run_command Nov 27 04:24:35 localhost ovn_controller[156436]: + CMD='/usr/bin/ovn-controller --pidfile unix:/run/openvswitch/db.sock ' Nov 27 04:24:35 localhost ovn_controller[156436]: + ARGS= Nov 27 04:24:35 localhost ovn_controller[156436]: + sudo kolla_copy_cacerts Nov 27 04:24:35 localhost systemd[1]: Started Session c13 of User root. Nov 27 04:24:35 localhost ovn_controller[156436]: + [[ ! -n '' ]] Nov 27 04:24:35 localhost ovn_controller[156436]: + . kolla_extend_start Nov 27 04:24:35 localhost ovn_controller[156436]: Running command: '/usr/bin/ovn-controller --pidfile unix:/run/openvswitch/db.sock ' Nov 27 04:24:35 localhost ovn_controller[156436]: + echo 'Running command: '\''/usr/bin/ovn-controller --pidfile unix:/run/openvswitch/db.sock '\''' Nov 27 04:24:35 localhost systemd[1]: session-c13.scope: Deactivated successfully. Nov 27 04:24:35 localhost ovn_controller[156436]: + umask 0022 Nov 27 04:24:35 localhost ovn_controller[156436]: + exec /usr/bin/ovn-controller --pidfile unix:/run/openvswitch/db.sock Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00001|reconnect|INFO|unix:/run/openvswitch/db.sock: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00002|reconnect|INFO|unix:/run/openvswitch/db.sock: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00003|main|INFO|OVN internal version is : [24.03.7-20.33.0-76.8] Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00004|main|INFO|OVS IDL reconnected, force recompute. Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00005|reconnect|INFO|tcp:ovsdbserver-sb.openstack.svc:6642: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00006|main|INFO|OVNSB IDL reconnected, force recompute. Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00007|reconnect|INFO|tcp:ovsdbserver-sb.openstack.svc:6642: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00008|features|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting to switch Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00009|rconn|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00010|rconn|INFO|unix:/var/run/openvswitch/br-int.mgmt: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00011|features|INFO|OVS Feature: ct_zero_snat, state: supported Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00012|features|INFO|OVS Feature: ct_flush, state: supported Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00013|reconnect|INFO|unix:/run/openvswitch/db.sock: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00014|main|INFO|OVS feature set changed, force recompute. Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00015|ofctrl|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting to switch Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00016|rconn|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00017|ovn_bfd|INFO|Disabled BFD on interface ovn-eb0c39-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00018|ovn_bfd|INFO|Disabled BFD on interface ovn-90a3af-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00019|ovn_bfd|INFO|Disabled BFD on interface ovn-a848a6-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00020|rconn|INFO|unix:/var/run/openvswitch/br-int.mgmt: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00021|ofctrl|INFO|ofctrl-wait-before-clear is now 8000 ms (was 0 ms) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00022|main|INFO|OVS OpenFlow connection reconnected,force recompute. Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00023|reconnect|INFO|unix:/run/openvswitch/db.sock: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00024|features|INFO|OVS DB schema supports 4 flow table prefixes, our IDL supports: 4 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00025|binding|INFO|Claiming lport a2718872-3b82-44ca-8d67-a0e516b2708b for this chassis. Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00026|binding|INFO|a2718872-3b82-44ca-8d67-a0e516b2708b: Claiming fa:16:3e:86:b9:b3 192.168.0.212 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00027|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00001|pinctrl(ovn_pinctrl0)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting to switch Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00028|binding|INFO|Removing lport a2718872-3b82-44ca-8d67-a0e516b2708b ovn-installed in OVS Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00002|rconn(ovn_pinctrl0)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00001|statctrl(ovn_statctrl3)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting to switch Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00003|rconn(ovn_pinctrl0)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00002|rconn(ovn_statctrl3)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connecting... Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00003|rconn(ovn_statctrl3)|INFO|unix:/var/run/openvswitch/br-int.mgmt: connected Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00029|ovn_bfd|INFO|Enabled BFD on interface ovn-eb0c39-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00030|ovn_bfd|INFO|Enabled BFD on interface ovn-90a3af-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00031|ovn_bfd|INFO|Enabled BFD on interface ovn-a848a6-0 Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00032|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00033|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00034|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00035|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00036|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:35 localhost ovn_controller[156436]: 2025-11-27T09:24:35Z|00037|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:36 localhost ovn_controller[156436]: 2025-11-27T09:24:36Z|00038|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:36 localhost ovn_controller[156436]: 2025-11-27T09:24:36Z|00039|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:37 localhost python3.9[156641]: ansible-ansible.legacy.command Invoked with _raw_params=ovs-vsctl remove open . other_config hw-offload#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:24:37 localhost ovs-vsctl[156642]: ovs|00001|vsctl|INFO|Called as ovs-vsctl remove open . other_config hw-offload Nov 27 04:24:37 localhost ovn_controller[156436]: 2025-11-27T09:24:37Z|00040|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:37 localhost ovn_controller[156436]: 2025-11-27T09:24:37Z|00041|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:24:37 localhost python3.9[156734]: ansible-ansible.legacy.command Invoked with _raw_params=ovs-vsctl get Open_vSwitch . external_ids:ovn-cms-options | sed 's/\"//g'#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:24:37 localhost ovs-vsctl[156736]: ovs|00001|db_ctl_base|ERR|no key "ovn-cms-options" in Open_vSwitch record "." column external_ids Nov 27 04:24:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46128 DF PROTO=TCP SPT=59040 DPT=9101 SEQ=3176879867 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B61DA0000000001030307) Nov 27 04:24:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12432 DF PROTO=TCP SPT=41394 DPT=9882 SEQ=3651847113 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B63B70000000001030307) Nov 27 04:24:39 localhost python3.9[156829]: ansible-ansible.legacy.command Invoked with _raw_params=ovs-vsctl remove Open_vSwitch . external_ids ovn-cms-options#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:24:39 localhost ovs-vsctl[156830]: ovs|00001|vsctl|INFO|Called as ovs-vsctl remove Open_vSwitch . external_ids ovn-cms-options Nov 27 04:24:39 localhost systemd[1]: session-50.scope: Deactivated successfully. Nov 27 04:24:39 localhost systemd[1]: session-50.scope: Consumed 40.920s CPU time. Nov 27 04:24:39 localhost systemd-logind[761]: Session 50 logged out. Waiting for processes to exit. Nov 27 04:24:39 localhost systemd-logind[761]: Removed session 50. Nov 27 04:24:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12434 DF PROTO=TCP SPT=41394 DPT=9882 SEQ=3651847113 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B6FDB0000000001030307) Nov 27 04:24:43 localhost ovn_controller[156436]: 2025-11-27T09:24:43Z|00042|binding|INFO|Setting lport a2718872-3b82-44ca-8d67-a0e516b2708b ovn-installed in OVS Nov 27 04:24:43 localhost ovn_controller[156436]: 2025-11-27T09:24:43Z|00043|binding|INFO|Setting lport a2718872-3b82-44ca-8d67-a0e516b2708b up in Southbound Nov 27 04:24:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59512 DF PROTO=TCP SPT=59846 DPT=9102 SEQ=1682526095 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B7E480000000001030307) Nov 27 04:24:45 localhost systemd[1]: Stopping User Manager for UID 0... Nov 27 04:24:45 localhost systemd[156466]: Activating special unit Exit the Session... Nov 27 04:24:45 localhost systemd[156466]: Stopped target Main User Target. Nov 27 04:24:45 localhost systemd[156466]: Stopped target Basic System. Nov 27 04:24:45 localhost systemd[156466]: Stopped target Paths. Nov 27 04:24:45 localhost systemd[156466]: Stopped target Sockets. Nov 27 04:24:45 localhost systemd[156466]: Stopped target Timers. Nov 27 04:24:45 localhost systemd[156466]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 04:24:45 localhost systemd[156466]: Closed D-Bus User Message Bus Socket. Nov 27 04:24:45 localhost systemd[156466]: Stopped Create User's Volatile Files and Directories. Nov 27 04:24:45 localhost systemd[156466]: Removed slice User Application Slice. Nov 27 04:24:45 localhost systemd[156466]: Reached target Shutdown. Nov 27 04:24:45 localhost systemd[156466]: Finished Exit the Session. Nov 27 04:24:45 localhost systemd[156466]: Reached target Exit the Session. Nov 27 04:24:45 localhost systemd[1]: user@0.service: Deactivated successfully. Nov 27 04:24:45 localhost systemd[1]: Stopped User Manager for UID 0. Nov 27 04:24:45 localhost systemd[1]: Stopping User Runtime Directory /run/user/0... Nov 27 04:24:45 localhost systemd[1]: run-user-0.mount: Deactivated successfully. Nov 27 04:24:45 localhost systemd[1]: user-runtime-dir@0.service: Deactivated successfully. Nov 27 04:24:45 localhost systemd[1]: Stopped User Runtime Directory /run/user/0. Nov 27 04:24:45 localhost systemd[1]: Removed slice User Slice of UID 0. Nov 27 04:24:46 localhost sshd[156849]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:24:46 localhost systemd-logind[761]: New session 52 of user zuul. Nov 27 04:24:46 localhost systemd[1]: Started Session 52 of User zuul. Nov 27 04:24:47 localhost python3.9[156942]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:24:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49580 DF PROTO=TCP SPT=39686 DPT=9100 SEQ=2072107359 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B865A0000000001030307) Nov 27 04:24:48 localhost python3.9[157038]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:49 localhost python3.9[157130]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/neutron setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:49 localhost python3.9[157222]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/neutron/kill_scripts setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:50 localhost python3.9[157314]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/neutron/ovn-metadata-proxy setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:50 localhost python3.9[157406]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/neutron/external/pids setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:51 localhost python3.9[157496]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'selinux'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:24:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49581 DF PROTO=TCP SPT=39686 DPT=9100 SEQ=2072107359 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B961A0000000001030307) Nov 27 04:24:52 localhost python3.9[157588]: ansible-ansible.posix.seboolean Invoked with name=virt_sandbox_use_netlink persistent=True state=True ignore_selinux_state=False Nov 27 04:24:53 localhost python3.9[157678]: ansible-ansible.legacy.stat Invoked with path=/var/lib/neutron/ovn_metadata_haproxy_wrapper follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53173 DF PROTO=TCP SPT=34326 DPT=9101 SEQ=1868278541 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128B9E1A0000000001030307) Nov 27 04:24:53 localhost python3.9[157751]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/neutron/ovn_metadata_haproxy_wrapper mode=0755 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235492.6892898-221-67761574977518/.source follow=False _original_basename=haproxy.j2 checksum=95c62e64c8f82dd9393a560d1b052dc98d38f810 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:55 localhost python3.9[157842]: ansible-ansible.legacy.stat Invoked with path=/var/lib/neutron/kill_scripts/haproxy-kill follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:24:55 localhost python3.9[157915]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/neutron/kill_scripts/haproxy-kill mode=0755 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235494.6876984-266-5520142465698/.source follow=False _original_basename=kill-script.j2 checksum=2dfb5489f491f61b95691c3bf95fa1fe48ff3700 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:24:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47071 DF PROTO=TCP SPT=58974 DPT=9101 SEQ=2634355822 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BA9DA0000000001030307) Nov 27 04:24:57 localhost python3.9[158007]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:24:57 localhost python3.9[158061]: ansible-ansible.legacy.dnf Invoked with name=['openvswitch3.3'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:24:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=63431 DF PROTO=TCP SPT=42360 DPT=9105 SEQ=3193687268 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BB5DA0000000001030307) Nov 27 04:25:02 localhost python3.9[158155]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=openvswitch.service state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:25:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=3456 DF PROTO=TCP SPT=60552 DPT=9105 SEQ=1626599328 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BC21A0000000001030307) Nov 27 04:25:03 localhost python3.9[158248]: ansible-ansible.legacy.stat Invoked with path=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/01-rootwrap.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:04 localhost python3.9[158319]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/01-rootwrap.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235503.4095614-376-169498526064317/.source.conf follow=False _original_basename=rootwrap.conf.j2 checksum=11f2cfb4b7d97b2cef3c2c2d88089e6999cffe22 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:04 localhost python3.9[158439]: ansible-ansible.legacy.stat Invoked with path=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/01-neutron-ovn-metadata-agent.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:25:06 localhost systemd[1]: tmp-crun.R6sXvk.mount: Deactivated successfully. Nov 27 04:25:06 localhost podman[158542]: 2025-11-27 09:25:06.009747148 +0000 UTC m=+0.098156328 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=starting, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_controller, org.label-schema.schema-version=1.0) Nov 27 04:25:06 localhost ovn_controller[156436]: 2025-11-27T09:25:06Z|00044|memory|INFO|17116 kB peak resident set size after 30.3 seconds Nov 27 04:25:06 localhost ovn_controller[156436]: 2025-11-27T09:25:06Z|00045|memory|INFO|idl-cells-OVN_Southbound:4033 idl-cells-Open_vSwitch:1045 if_status_mgr_ifaces_state_usage-KB:1 if_status_mgr_ifaces_usage-KB:1 lflow-cache-entries-cache-expr:76 lflow-cache-entries-cache-matches:195 lflow-cache-size-KB:289 local_datapath_usage-KB:1 ofctrl_desired_flow_usage-KB:154 ofctrl_installed_flow_usage-KB:112 ofctrl_sb_flow_ref_usage-KB:67 Nov 27 04:25:06 localhost python3.9[158541]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/01-neutron-ovn-metadata-agent.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235504.4857543-376-234153930663716/.source.conf follow=False _original_basename=neutron-ovn-metadata-agent.conf.j2 checksum=8bc979abbe81c2cf3993a225517a7e2483e20443 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:06 localhost podman[158542]: 2025-11-27 09:25:06.083168888 +0000 UTC m=+0.171578108 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, tcib_managed=true, container_name=ovn_controller) Nov 27 04:25:06 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:25:07 localhost python3.9[158671]: ansible-ansible.legacy.stat Invoked with path=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/10-neutron-metadata.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:07 localhost python3.9[158742]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/10-neutron-metadata.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235506.861821-508-137347552749998/.source.conf _original_basename=10-neutron-metadata.conf follow=False checksum=aa9e89725fbcebf7a5c773d7b97083445b7b7759 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53176 DF PROTO=TCP SPT=34326 DPT=9101 SEQ=1868278541 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BD5DB0000000001030307) Nov 27 04:25:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33806 DF PROTO=TCP SPT=39530 DPT=9882 SEQ=2483535418 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BD8E70000000001030307) Nov 27 04:25:08 localhost python3.9[158832]: ansible-ansible.legacy.stat Invoked with path=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/05-nova-metadata.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:09 localhost python3.9[158903]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent/05-nova-metadata.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235508.3606884-508-203961965496710/.source.conf _original_basename=05-nova-metadata.conf follow=False checksum=979187b925479d81d0609f4188e5b95fe1f92c18 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:09 localhost python3.9[158993]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:25:10 localhost python3.9[159087]: ansible-ansible.builtin.file Invoked with path=/var/local/libexec recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:11 localhost python3.9[159179]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-container-shutdown follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:11 localhost python3.9[159227]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-container-shutdown _original_basename=edpm-container-shutdown recurse=False state=file path=/var/local/libexec/edpm-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=33808 DF PROTO=TCP SPT=39530 DPT=9882 SEQ=2483535418 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BE4DA0000000001030307) Nov 27 04:25:12 localhost python3.9[159319]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-start-podman-container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:12 localhost python3.9[159367]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-start-podman-container _original_basename=edpm-start-podman-container recurse=False state=file path=/var/local/libexec/edpm-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:13 localhost python3.9[159459]: ansible-ansible.builtin.file Invoked with mode=420 path=/etc/systemd/system-preset state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:13 localhost ovn_controller[156436]: 2025-11-27T09:25:13Z|00046|memory_trim|INFO|Detected inactivity (last active 30004 ms ago): trimming memory Nov 27 04:25:14 localhost python3.9[159551]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm-container-shutdown.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:14 localhost python3.9[159599]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/edpm-container-shutdown.service _original_basename=edpm-container-shutdown-service recurse=False state=file path=/etc/systemd/system/edpm-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:15 localhost python3.9[159691]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:15 localhost python3.9[159739]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-edpm-container-shutdown.preset _original_basename=91-edpm-container-shutdown-preset recurse=False state=file path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23448 DF PROTO=TCP SPT=45008 DPT=9102 SEQ=1521064177 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BF37A0000000001030307) Nov 27 04:25:16 localhost python3.9[159831]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm-container-shutdown state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:25:16 localhost systemd[1]: Reloading. Nov 27 04:25:16 localhost systemd-rc-local-generator[159855]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:16 localhost systemd-sysv-generator[159861]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:16 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:17 localhost python3.9[159961]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/netns-placeholder.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13156 DF PROTO=TCP SPT=47608 DPT=9100 SEQ=3225916028 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128BFB9A0000000001030307) Nov 27 04:25:17 localhost python3.9[160009]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/etc/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:19 localhost python3.9[160101]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:19 localhost python3.9[160149]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/etc/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:20 localhost python3.9[160241]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=netns-placeholder state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:25:20 localhost systemd[1]: Reloading. Nov 27 04:25:20 localhost systemd-sysv-generator[160273]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:20 localhost systemd-rc-local-generator[160270]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:20 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:20 localhost systemd[1]: Starting Create netns directory... Nov 27 04:25:20 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 04:25:20 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 04:25:20 localhost systemd[1]: Finished Create netns directory. Nov 27 04:25:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13157 DF PROTO=TCP SPT=47608 DPT=9100 SEQ=3225916028 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C0B5B0000000001030307) Nov 27 04:25:23 localhost python3.9[160375]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/healthchecks setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8462 DF PROTO=TCP SPT=40270 DPT=9101 SEQ=2591844416 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C135B0000000001030307) Nov 27 04:25:24 localhost python3.9[160467]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/ovn_metadata_agent/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:25 localhost python3.9[160540]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/ovn_metadata_agent/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235524.0312903-961-30471892329399/.source _original_basename=healthcheck follow=False checksum=898a5a1fcd473cf731177fc866e3bd7ebf20a131 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:25 localhost python3.9[160632]: ansible-ansible.builtin.file Invoked with path=/var/lib/kolla/config_files recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:25:26 localhost python3.9[160724]: ansible-ansible.legacy.stat Invoked with path=/var/lib/kolla/config_files/ovn_metadata_agent.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:25:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7728 DF PROTO=TCP SPT=48246 DPT=9105 SEQ=476503843 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C1F5A0000000001030307) Nov 27 04:25:27 localhost python3.9[160799]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/kolla/config_files/ovn_metadata_agent.json mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235526.1724362-1036-121134719581422/.source.json _original_basename=.i0ytsnys follow=False checksum=a908ef151ded3a33ae6c9ac8be72a35e5e33b9dc backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:27 localhost python3.9[160891]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/edpm-config/container-startup-config/ovn_metadata_agent state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8464 DF PROTO=TCP SPT=40270 DPT=9101 SEQ=2591844416 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C2B1B0000000001030307) Nov 27 04:25:30 localhost python3.9[161148]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/edpm-config/container-startup-config/ovn_metadata_agent config_pattern=*.json debug=False Nov 27 04:25:31 localhost python3.9[161240]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:25:31 localhost python3.9[161332]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Nov 27 04:25:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7730 DF PROTO=TCP SPT=48246 DPT=9105 SEQ=476503843 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C371A0000000001030307) Nov 27 04:25:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:25:36 localhost podman[161451]: 2025-11-27 09:25:36.570422703 +0000 UTC m=+0.081565958 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:25:36 localhost podman[161451]: 2025-11-27 09:25:36.610860903 +0000 UTC m=+0.122004118 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 04:25:36 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:25:36 localhost python3[161450]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/edpm-config/container-startup-config/ovn_metadata_agent config_id=ovn_metadata_agent config_overrides={} config_patterns=*.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:25:36 localhost python3[161450]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "c64a92d8e8fa4f5fb5baf11a4a693a964be3868fb7e72462c6e612c604f8d071",#012 "Digest": "sha256:2b8255d3a22035616e569dbe22862a2560e15cdaefedae0059a354d558788e1e",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:2b8255d3a22035616e569dbe22862a2560e15cdaefedae0059a354d558788e1e"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:34:14.989876147Z",#012 "Config": {#012 "User": "neutron",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 784145152,#012 "VirtualSize": 784145152,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048/diff:/var/lib/containers/storage/overlay/47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35/diff:/var/lib/containers/storage/overlay/f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:c136b33417f134a3b932677bcf7a2df089c29f20eca250129eafd2132d4708bb",#012 "sha256:bc63f71478d9d90db803b468b28e5d9e0268adbace958b608ab10bd0819798bd",#012 "sha256:3277562ff4450bdcd859dd0b0be874b10dd6f3502be711d42aab9ff44a85cf28",#012 "sha256:982219792b3d83fa04ae12d0161dd3b982e7e3ed68293e6c876d50161b73746b"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "neutron",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf Nov 27 04:25:37 localhost podman[161528]: 2025-11-27 09:25:37.078481164 +0000 UTC m=+0.070817097 container remove 2753965e0806d98bb8365786da80da43cf572d39e847a2905a4bb8fc62609954 (image=registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1, name=ovn_metadata_agent, release=1761123044, tcib_managed=true, distribution-scope=public, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, com.redhat.component=openstack-neutron-metadata-agent-ovn-container, url=https://www.redhat.com, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-neutron-metadata-agent-ovn, build-date=2025-11-19T00:14:25Z, summary=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, io.buildah.version=1.41.4, vcs-ref=89d55f10f82ff50b4f24de36868d7c635c279c7c, name=rhosp17/openstack-neutron-metadata-agent-ovn, config_id=tripleo_step4, description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, managed_by=tripleo_ansible, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.expose-services=, version=17.1.12, org.opencontainers.image.revision=89d55f10f82ff50b4f24de36868d7c635c279c7c, konflux.additional-tags=17.1.12 17.1_20251118.1, maintainer=OpenStack TripleO Team, vendor=Red Hat, Inc., io.k8s.description=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, batch=17.1_20251118.1, io.k8s.display-name=Red Hat OpenStack Platform 17.1 neutron-metadata-agent-ovn, config_data={'cgroupns': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': '8cac2334de855b5d814608ea6483099b'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-neutron-metadata-agent-ovn:17.1', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'start_order': 1, 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/neutron:/var/log/neutron:z', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/neutron:/var/lib/kolla/config_files/src:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared,z', '/var/lib/neutron:/var/lib/neutron:shared,z', '/run/netns:/run/netns:shared', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro']}, container_name=ovn_metadata_agent) Nov 27 04:25:37 localhost python3[161450]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman rm --force ovn_metadata_agent Nov 27 04:25:37 localhost podman[161541]: Nov 27 04:25:37 localhost podman[161541]: 2025-11-27 09:25:37.175977474 +0000 UTC m=+0.081682122 container create 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 04:25:37 localhost podman[161541]: 2025-11-27 09:25:37.136237823 +0000 UTC m=+0.041942501 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 04:25:37 localhost python3[161450]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name ovn_metadata_agent --cgroupns=host --conmon-pidfile /run/ovn_metadata_agent.pid --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env EDPM_CONFIG_HASH=df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311 --healthcheck-command /openstack/healthcheck --label config_id=ovn_metadata_agent --label container_name=ovn_metadata_agent --label managed_by=edpm_ansible --label config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']} --log-driver journald --log-level info --network host --pid host --privileged=True --user root --volume /run/openvswitch:/run/openvswitch:z --volume /var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z --volume /run/netns:/run/netns:shared --volume /var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro --volume /var/lib/neutron:/var/lib/neutron:shared,z --volume /var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro --volume /var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro --volume /var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z --volume /var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 04:25:38 localhost python3.9[161670]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:25:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8465 DF PROTO=TCP SPT=40270 DPT=9101 SEQ=2591844416 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C4BDA0000000001030307) Nov 27 04:25:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4428 DF PROTO=TCP SPT=50488 DPT=9882 SEQ=234152369 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C4E170000000001030307) Nov 27 04:25:38 localhost python3.9[161764]: ansible-file Invoked with path=/etc/systemd/system/edpm_ovn_metadata_agent.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:39 localhost python3.9[161810]: ansible-stat Invoked with path=/etc/systemd/system/edpm_ovn_metadata_agent_healthcheck.timer follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:25:40 localhost python3.9[161901]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764235539.4219508-1300-13111965447329/source dest=/etc/systemd/system/edpm_ovn_metadata_agent.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:25:40 localhost python3.9[161947]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:25:40 localhost systemd[1]: Reloading. Nov 27 04:25:40 localhost systemd-rc-local-generator[161972]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:40 localhost systemd-sysv-generator[161977]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:41 localhost python3.9[162029]: ansible-systemd Invoked with state=restarted name=edpm_ovn_metadata_agent.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:25:41 localhost systemd[1]: Reloading. Nov 27 04:25:41 localhost systemd-rc-local-generator[162056]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:41 localhost systemd-sysv-generator[162061]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:41 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:41 localhost systemd[1]: Starting ovn_metadata_agent container... Nov 27 04:25:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4430 DF PROTO=TCP SPT=50488 DPT=9882 SEQ=234152369 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C5A1B0000000001030307) Nov 27 04:25:41 localhost systemd[1]: tmp-crun.df3SV3.mount: Deactivated successfully. Nov 27 04:25:41 localhost systemd[1]: Started libcrun container. Nov 27 04:25:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/cf0c71d487ffe8262f3245cd08432946824fea7e2b3684a48db754d7bef1589c/merged/etc/neutron.conf.d supports timestamps until 2038 (0x7fffffff) Nov 27 04:25:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/cf0c71d487ffe8262f3245cd08432946824fea7e2b3684a48db754d7bef1589c/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 04:25:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:25:41 localhost podman[162072]: 2025-11-27 09:25:41.950864534 +0000 UTC m=+0.131539106 container init 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:25:41 localhost ovn_metadata_agent[162087]: + sudo -E kolla_set_configs Nov 27 04:25:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:25:41 localhost podman[162072]: 2025-11-27 09:25:41.978711011 +0000 UTC m=+0.159385623 container start 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent) Nov 27 04:25:41 localhost edpm-start-podman-container[162072]: ovn_metadata_agent Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Validating config file Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Copying service configuration files Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Deleting /etc/neutron/rootwrap.conf Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Copying /etc/neutron.conf.d/01-rootwrap.conf to /etc/neutron/rootwrap.conf Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /etc/neutron/rootwrap.conf Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Writing out command to execute Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/kill_scripts Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/.cache Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/external Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/ovn-metadata-proxy Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/ovn_metadata_haproxy_wrapper Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/metadata_proxy Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/kill_scripts/haproxy-kill Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/.cache/python-entrypoints Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/.cache/python-entrypoints/adac9f827fd7fb11fb07020ef60ee06a1fede4feab743856dc8fb3266181d934 Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/external/pids Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: INFO:__main__:Setting permission for /var/lib/neutron/ovn-metadata-proxy/3d77aad9-8375-4ff0-8227-0972f0af9ab9.conf Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: ++ cat /run_command Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + CMD=neutron-ovn-metadata-agent Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + ARGS= Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + sudo kolla_copy_cacerts Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: Running command: 'neutron-ovn-metadata-agent' Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + [[ ! -n '' ]] Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + . kolla_extend_start Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + echo 'Running command: '\''neutron-ovn-metadata-agent'\''' Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + umask 0022 Nov 27 04:25:42 localhost ovn_metadata_agent[162087]: + exec neutron-ovn-metadata-agent Nov 27 04:25:42 localhost edpm-start-podman-container[162071]: Creating additional drop-in dependency for "ovn_metadata_agent" (1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51) Nov 27 04:25:42 localhost systemd[1]: Reloading. Nov 27 04:25:42 localhost podman[162094]: 2025-11-27 09:25:42.12947036 +0000 UTC m=+0.141892559 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=starting, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 04:25:42 localhost podman[162094]: 2025-11-27 09:25:42.137739604 +0000 UTC m=+0.150161803 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:25:42 localhost systemd-rc-local-generator[162161]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:42 localhost systemd-sysv-generator[162166]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:42 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:42 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:25:42 localhost systemd[1]: Started ovn_metadata_agent container. Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.498 162092 INFO neutron.common.config [-] Logging enabled!#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.498 162092 INFO neutron.common.config [-] /usr/bin/neutron-ovn-metadata-agent version 22.2.2.dev43#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.498 162092 DEBUG neutron.common.config [-] command line: /usr/bin/neutron-ovn-metadata-agent setup_logging /usr/lib/python3.9/site-packages/neutron/common/config.py:123#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.498 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] command line args: [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] config files: ['/etc/neutron/neutron.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] agent_down_time = 75 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] allow_bulk = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] api_extensions_path = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] api_paste_config = api-paste.ini log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.499 162092 DEBUG neutron.agent.ovn.metadata_agent [-] api_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] auth_ca_cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] auth_strategy = keystone log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] backlog = 4096 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] base_mac = fa:16:3e:00:00:00 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] bind_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] bind_port = 9696 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] client_socket_timeout = 900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] config_dir = ['/etc/neutron.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] config_file = ['/etc/neutron/neutron.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.500 162092 DEBUG neutron.agent.ovn.metadata_agent [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] control_exchange = neutron log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] core_plugin = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] default_availability_zones = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'OFPHandler=INFO', 'OfctlService=INFO', 'os_ken.base.app_manager=INFO', 'os_ken.controller.controller=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] dhcp_agent_notification = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] dhcp_lease_duration = 86400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.501 162092 DEBUG neutron.agent.ovn.metadata_agent [-] dhcp_load_type = networks log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] dns_domain = openstacklocal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] enable_new_agents = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] enable_traditional_dhcp = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] external_dns_driver = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] external_pids = /var/lib/neutron/external/pids log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] filter_validation = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] global_physnet_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.502 162092 DEBUG neutron.agent.ovn.metadata_agent [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ipam_driver = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ipv6_pd_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.503 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] max_dns_nameservers = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] max_header_line = 16384 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.504 162092 DEBUG neutron.agent.ovn.metadata_agent [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] max_subnet_host_routes = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_backlog = 4096 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_proxy_group = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_proxy_shared_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_proxy_socket = /var/lib/neutron/metadata_proxy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_proxy_socket_mode = deduce log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_proxy_user = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.505 162092 DEBUG neutron.agent.ovn.metadata_agent [-] metadata_workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] network_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] notify_nova_on_port_data_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] notify_nova_on_port_status_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_client_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_client_priv_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_metadata_host = nova-metadata-internal.openstack.svc log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_metadata_insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_metadata_port = 8775 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.506 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova_metadata_protocol = http log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] pagination_max_limit = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] periodic_fuzzy_delay = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] periodic_interval = 40 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] retry_until_window = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rpc_resources_processing_step = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.507 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rpc_response_max_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rpc_state_report_workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] rpc_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] send_events_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] service_plugins = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] setproctitle = on log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] state_path = /var/lib/neutron log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] syslog_log_facility = syslog log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] tcp_keepidle = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.508 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] vlan_transparent = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] wsgi_default_pool_size = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] wsgi_keep_alive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.509 162092 DEBUG neutron.agent.ovn.metadata_agent [-] wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] wsgi_server_debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_concurrency.disable_process_locking = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_concurrency.lock_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.connection_string = messaging:// log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.es_doc_type = notification log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.es_scroll_size = 10000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.es_scroll_time = 2m log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.510 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.filter_error_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.hmac_keys = SECRET_KEY log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.sentinel_service_name = mymaster log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.socket_timeout = 0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] profiler.trace_sqlalchemy = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.enforce_new_defaults = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.enforce_scope = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.policy_default_rule = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.policy_dirs = ['policy.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.511 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.policy_file = policy.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.remote_content_type = application/x-www-form-urlencoded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.remote_ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.remote_ssl_client_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.remote_ssl_client_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_policy.remote_ssl_verify_server_crt = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_metrics.metrics_buffer_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_metrics.metrics_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_metrics.metrics_process_name = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.512 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] service_providers.service_provider = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.capabilities = [21, 12, 1, 2, 19] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.capabilities = [21, 12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.513 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_dhcp_release.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.capabilities = [21, 12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.514 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_ovs_vsctl.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.capabilities = [21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_namespace.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.capabilities = [12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.515 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_conntrack.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.capabilities = [12, 21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] privsep_link.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.check_child_processes_action = respawn log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.516 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.check_child_processes_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.comment_iptables_rules = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.debug_iptables_rules = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.kill_scripts_path = /etc/neutron/kill_scripts/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.root_helper = sudo neutron-rootwrap /etc/neutron/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.root_helper_daemon = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.use_helper_for_ns_read = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] AGENT.use_random_fully = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_versionedobjects.fatal_exception_format_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.517 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.default_quota = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_driver = neutron.db.quota.driver_nolock.DbQuotaNoLockDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_network = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_port = 500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_security_group = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_security_group_rule = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.quota_subnet = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] QUOTAS.track_quota_usage = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.518 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.endpoint_type = public log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] nova.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.519 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.endpoint_type = public log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] placement.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.520 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.enable_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.521 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.service_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.522 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.valid_interfaces = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ironic.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] cli_script.dry_run = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.allow_stateless_action_supported = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.dhcp_default_lease_time = 43200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.disable_ovn_dhcp_for_baremetal_ports = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.dns_servers = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.enable_distributed_floating_ip = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.neutron_sync_mode = log log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.523 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_dhcp4_global_options = {} log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_dhcp6_global_options = {} log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_emit_need_to_frag = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_l3_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_l3_scheduler = leastloaded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_metadata_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_nb_ca_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_nb_certificate = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_nb_connection = tcp:127.0.0.1:6641 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.524 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_nb_private_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_sb_ca_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_sb_certificate = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_sb_connection = tcp:ovsdbserver-sb.openstack.svc:6642 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovn_sb_private_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovsdb_connection_timeout = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovsdb_log_level = INFO log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovsdb_probe_interval = 60000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.ovsdb_retry_max_interval = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.vhost_sock_dir = /var/run/openvswitch log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.525 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovn.vif_type = ovs log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] OVS.bridge_mac_table_size = 50000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] OVS.igmp_snooping_enable = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] OVS.ovsdb_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovs.ovsdb_connection = tcp:127.0.0.1:6640 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ovs.ovsdb_connection_timeout = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.amqp_durable_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.526 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.527 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_quorum_queue = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.528 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_notifications.driver = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.529 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.530 162092 DEBUG neutron.agent.ovn.metadata_agent [-] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.530 162092 DEBUG neutron.agent.ovn.metadata_agent [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.538 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Bridge.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.538 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Port.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.538 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Interface.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.538 162092 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: connecting...#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.539 162092 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: connected#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.554 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Loaded chassis name 927ada8e-9e05-4f20-b8c8-d7c6f26f86ee (UUID: 927ada8e-9e05-4f20-b8c8-d7c6f26f86ee) and ovn bridge br-int. _load_config /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:309#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.576 162092 INFO neutron.agent.ovn.metadata.ovsdb [-] Getting OvsdbSbOvnIdl for MetadataAgent with retry#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.576 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created lookup_table index Chassis.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:87#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.576 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Datapath_Binding.tunnel_key autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.576 162092 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Chassis_Private.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.578 162092 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:ovsdbserver-sb.openstack.svc:6642: connecting...#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.579 162092 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:ovsdbserver-sb.openstack.svc:6642: connected#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.586 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched CREATE: PortBindingCreateWithChassis(events=('create',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:86:b9:b3 192.168.0.212'], port_security=['fa:16:3e:86:b9:b3 192.168.0.212'], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '192.168.0.212/24', 'neutron:device_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'neutron:device_owner': 'compute:nova', 'neutron:host_id': 'np0005537446.localdomain', 'neutron:mtu': '', 'neutron:network_name': 'neutron-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'neutron:port_capabilities': '', 'neutron:port_fip': '192.168.122.20', 'neutron:port_name': '', 'neutron:project_id': '02a308de13ca4518beaddd402ba93936', 'neutron:revision_number': '7', 'neutron:security_group_ids': '05e0d7b2-5469-4d02-a67e-d8080e0339fc 869e5a88-209d-40d4-906d-1e63476c9c16', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=f32dcab3-a58b-44c3-83ba-19ec8eecda08, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=a2718872-3b82-44ca-8d67-a0e516b2708b) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.587 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched CREATE: ChassisPrivateCreateEvent(events=('create',), table='Chassis_Private', conditions=(('name', '=', '927ada8e-9e05-4f20-b8c8-d7c6f26f86ee'),), old_conditions=None), priority=20 to row=Chassis_Private(chassis=[], external_ids={'neutron:ovn-metadata-id': '42f03390-922a-5d0c-a38e-2abedc60cd10', 'neutron:ovn-metadata-sb-cfg': '1'}, name=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, nb_cfg_timestamp=1764235484169, nb_cfg=4) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.587 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a2718872-3b82-44ca-8d67-a0e516b2708b in datapath 3d77aad9-8375-4ff0-8227-0972f0af9ab9 bound to our chassis on insert#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.588 162092 DEBUG neutron_lib.callbacks.manager [-] Subscribe: > process after_init 55550000, False subscribe /usr/lib/python3.9/site-packages/neutron_lib/callbacks/manager.py:52#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.588 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.588 162092 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.589 162092 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.589 162092 INFO oslo_service.service [-] Starting 1 workers#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.591 162092 DEBUG oslo_service.service [-] Started child 162191 _start_child /usr/lib/python3.9/site-packages/oslo_service/service.py:575#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.593 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network 3d77aad9-8375-4ff0-8227-0972f0af9ab9#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.594 162092 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/neutron/neutron.conf', '--config-dir', '/etc/neutron.conf.d', '--privsep_context', 'neutron.privileged.default', '--privsep_sock_path', '/tmp/tmprzhyq439/privsep.sock']#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.596 162191 DEBUG neutron_lib.callbacks.manager [-] Publish callbacks ['neutron.agent.ovn.metadata.server.MetadataProxyHandler.post_fork_initialize-392349'] for process (None), after_init _notify_loop /usr/lib/python3.9/site-packages/neutron_lib/callbacks/manager.py:184#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.622 162191 INFO neutron.agent.ovn.metadata.ovsdb [-] Getting OvsdbSbOvnIdl for MetadataAgent with retry#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.623 162191 DEBUG ovsdbapp.backend.ovs_idl [-] Created lookup_table index Chassis.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:87#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.623 162191 DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Datapath_Binding.tunnel_key autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.631 162191 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:ovsdbserver-sb.openstack.svc:6642: connecting...#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.633 162191 INFO ovsdbapp.backend.ovs_idl.vlog [-] tcp:ovsdbserver-sb.openstack.svc:6642: connected#033[00m Nov 27 04:25:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:43.650 162191 INFO eventlet.wsgi.server [-] (162191) wsgi starting up on http:/var/lib/neutron/metadata_proxy#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.171 162092 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.172 162092 DEBUG oslo.privsep.daemon [-] Accepted privsep connection to /tmp/tmprzhyq439/privsep.sock __init__ /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:362#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.077 162196 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.082 162196 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.086 162196 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_DAC_OVERRIDE|CAP_DAC_READ_SEARCH|CAP_NET_ADMIN|CAP_SYS_ADMIN|CAP_SYS_PTRACE/CAP_DAC_OVERRIDE|CAP_DAC_READ_SEARCH|CAP_NET_ADMIN|CAP_SYS_ADMIN|CAP_SYS_PTRACE/none#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.086 162196 INFO oslo.privsep.daemon [-] privsep daemon running as pid 162196#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.176 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[6b57109f-c947-4131-a2dd-eec9bf7d866d]: (2,) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:44 localhost systemd[1]: session-52.scope: Deactivated successfully. Nov 27 04:25:44 localhost systemd[1]: session-52.scope: Consumed 31.685s CPU time. Nov 27 04:25:44 localhost systemd-logind[761]: Session 52 logged out. Waiting for processes to exit. Nov 27 04:25:44 localhost systemd-logind[761]: Removed session 52. Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.588 162196 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "context-manager" by "neutron_lib.db.api._create_context_manager" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.588 162196 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" acquired by "neutron_lib.db.api._create_context_manager" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:25:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:44.588 162196 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" "released" by "neutron_lib.db.api._create_context_manager" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.035 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b142ba5a-cc8f-4d7f-b343-14c13556cf58]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.037 162092 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/neutron/neutron.conf', '--config-dir', '/etc/neutron.conf.d', '--privsep_context', 'neutron.privileged.link_cmd', '--privsep_sock_path', '/tmp/tmp3os5wls2/privsep.sock']#033[00m Nov 27 04:25:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39540 DF PROTO=TCP SPT=41238 DPT=9102 SEQ=434664775 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C68A90000000001030307) Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.650 162092 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.650 162092 DEBUG oslo.privsep.daemon [-] Accepted privsep connection to /tmp/tmp3os5wls2/privsep.sock __init__ /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:362#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.547 162207 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.552 162207 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.556 162207 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_NET_ADMIN|CAP_SYS_ADMIN/CAP_NET_ADMIN|CAP_SYS_ADMIN/none#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.556 162207 INFO oslo.privsep.daemon [-] privsep daemon running as pid 162207#033[00m Nov 27 04:25:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:45.653 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[8a4f0d19-9119-4878-b25a-733303a7ea6c]: (2,) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.084 162207 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "context-manager" by "neutron_lib.db.api._create_context_manager" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.084 162207 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" acquired by "neutron_lib.db.api._create_context_manager" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.084 162207 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" "released" by "neutron_lib.db.api._create_context_manager" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.541 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[0b036b2e-d2ad-47bf-a7ec-24c1b7f15869]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.543 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[72b41e21-a0e0-492b-94df-ccdb909b1fee]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.568 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[4f3b2381-2164-44ab-b606-e99b2e6a3298]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.579 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[cab646c7-1eed-456a-bd7c-e780e094d0f7]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap3d77aad9-81'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:76:bb:fc'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 104, 'tx_packets': 68, 'rx_bytes': 8926, 'tx_bytes': 7143, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 104, 'tx_packets': 68, 'rx_bytes': 8926, 'tx_bytes': 7143, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 14], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483664], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 662132, 'reachable_time': 33238, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 6, 'inoctets': 448, 'indelivers': 1, 'outforwdatagrams': 0, 'outpkts': 17, 'outoctets': 1164, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 6, 'outmcastpkts': 17, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 448, 'outmcastoctets': 1164, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 6, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 1, 'inerrors': 0, 'outmsgs': 17, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 162217, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.594 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ced7a3fb-e18c-4d5c-ac33-b6ed9667279f]: (4, ({'family': 2, 'prefixlen': 32, 'flags': 128, 'scope': 0, 'index': 2, 'attrs': [['IFA_ADDRESS', '169.254.169.254'], ['IFA_LOCAL', '169.254.169.254'], ['IFA_BROADCAST', '169.254.169.254'], ['IFA_LABEL', 'tap3d77aad9-81'], ['IFA_FLAGS', 128], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 662138, 'tstamp': 662138}]], 'header': {'length': 96, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 162218, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'}, {'family': 2, 'prefixlen': 24, 'flags': 128, 'scope': 0, 'index': 2, 'attrs': [['IFA_ADDRESS', '192.168.0.2'], ['IFA_LOCAL', '192.168.0.2'], ['IFA_BROADCAST', '192.168.0.255'], ['IFA_LABEL', 'tap3d77aad9-81'], ['IFA_FLAGS', 128], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 662143, 'tstamp': 662143}]], 'header': {'length': 96, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 162218, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'}, {'family': 10, 'prefixlen': 64, 'flags': 128, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::a9fe:a9fe'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 662140, 'tstamp': 662140}], ['IFA_FLAGS', 128]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 162218, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'}, {'family': 10, 'prefixlen': 64, 'flags': 128, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:fe76:bbfc'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 662132, 'tstamp': 662132}], ['IFA_FLAGS', 128]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 162218, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'})) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.652 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[7cc54a3f-959c-4383-8ebc-a48e6e456c53]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.654 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap3d77aad9-80, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.658 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap3d77aad9-80, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.658 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.659 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tap3d77aad9-80, col_values=(('external_ids', {'iface-id': '60750ee5-275f-4987-8dcc-15b61f3cb86a'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.659 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:25:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:46.663 162092 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/neutron/neutron.conf', '--config-dir', '/etc/neutron.conf.d', '--privsep_context', 'neutron.privileged.namespace_cmd', '--privsep_sock_path', '/tmp/tmptta4bux2/privsep.sock']#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.246 162092 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.247 162092 DEBUG oslo.privsep.daemon [-] Accepted privsep connection to /tmp/tmptta4bux2/privsep.sock __init__ /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:362#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.141 162227 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.147 162227 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.150 162227 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_SYS_ADMIN/CAP_SYS_ADMIN/none#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.151 162227 INFO oslo.privsep.daemon [-] privsep daemon running as pid 162227#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.250 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[d40a14ad-1896-47cf-9807-c98d04edbbf4]: (2,) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.652 162227 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "context-manager" by "neutron_lib.db.api._create_context_manager" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.652 162227 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" acquired by "neutron_lib.db.api._create_context_manager" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:25:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:47.653 162227 DEBUG oslo_concurrency.lockutils [-] Lock "context-manager" "released" by "neutron_lib.db.api._create_context_manager" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:25:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18443 DF PROTO=TCP SPT=52324 DPT=9100 SEQ=1760260511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C70DA0000000001030307) Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.089 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[2b9293f0-59c0-462d-b27b-9fcfc97276c9]: (4, ['ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9']) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.092 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbAddCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, column=external_ids, values=({'neutron:ovn-metadata-id': '42f03390-922a-5d0c-a38e-2abedc60cd10'},)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.093 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.094 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-bridge': 'br-int'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.120 162092 DEBUG oslo_service.service [-] Full set of CONF: wait /usr/lib/python3.9/site-packages/oslo_service/service.py:649#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.120 162092 DEBUG oslo_service.service [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.121 162092 DEBUG oslo_service.service [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.121 162092 DEBUG oslo_service.service [-] command line args: [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.121 162092 DEBUG oslo_service.service [-] config files: ['/etc/neutron/neutron.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.121 162092 DEBUG oslo_service.service [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.121 162092 DEBUG oslo_service.service [-] agent_down_time = 75 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.122 162092 DEBUG oslo_service.service [-] allow_bulk = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.122 162092 DEBUG oslo_service.service [-] api_extensions_path = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.122 162092 DEBUG oslo_service.service [-] api_paste_config = api-paste.ini log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.122 162092 DEBUG oslo_service.service [-] api_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.123 162092 DEBUG oslo_service.service [-] auth_ca_cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.123 162092 DEBUG oslo_service.service [-] auth_strategy = keystone log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.123 162092 DEBUG oslo_service.service [-] backlog = 4096 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.123 162092 DEBUG oslo_service.service [-] base_mac = fa:16:3e:00:00:00 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.124 162092 DEBUG oslo_service.service [-] bind_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.124 162092 DEBUG oslo_service.service [-] bind_port = 9696 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.124 162092 DEBUG oslo_service.service [-] client_socket_timeout = 900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.124 162092 DEBUG oslo_service.service [-] config_dir = ['/etc/neutron.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.125 162092 DEBUG oslo_service.service [-] config_file = ['/etc/neutron/neutron.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.125 162092 DEBUG oslo_service.service [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.125 162092 DEBUG oslo_service.service [-] control_exchange = neutron log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.125 162092 DEBUG oslo_service.service [-] core_plugin = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.126 162092 DEBUG oslo_service.service [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.126 162092 DEBUG oslo_service.service [-] default_availability_zones = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.126 162092 DEBUG oslo_service.service [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'OFPHandler=INFO', 'OfctlService=INFO', 'os_ken.base.app_manager=INFO', 'os_ken.controller.controller=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.126 162092 DEBUG oslo_service.service [-] dhcp_agent_notification = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.127 162092 DEBUG oslo_service.service [-] dhcp_lease_duration = 86400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.127 162092 DEBUG oslo_service.service [-] dhcp_load_type = networks log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.127 162092 DEBUG oslo_service.service [-] dns_domain = openstacklocal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.127 162092 DEBUG oslo_service.service [-] enable_new_agents = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.128 162092 DEBUG oslo_service.service [-] enable_traditional_dhcp = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.128 162092 DEBUG oslo_service.service [-] external_dns_driver = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.128 162092 DEBUG oslo_service.service [-] external_pids = /var/lib/neutron/external/pids log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.128 162092 DEBUG oslo_service.service [-] filter_validation = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.128 162092 DEBUG oslo_service.service [-] global_physnet_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.129 162092 DEBUG oslo_service.service [-] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.129 162092 DEBUG oslo_service.service [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.129 162092 DEBUG oslo_service.service [-] http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.130 162092 DEBUG oslo_service.service [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.130 162092 DEBUG oslo_service.service [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.130 162092 DEBUG oslo_service.service [-] ipam_driver = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.130 162092 DEBUG oslo_service.service [-] ipv6_pd_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.130 162092 DEBUG oslo_service.service [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.131 162092 DEBUG oslo_service.service [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.131 162092 DEBUG oslo_service.service [-] log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.131 162092 DEBUG oslo_service.service [-] log_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.131 162092 DEBUG oslo_service.service [-] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.131 162092 DEBUG oslo_service.service [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.132 162092 DEBUG oslo_service.service [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.132 162092 DEBUG oslo_service.service [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.132 162092 DEBUG oslo_service.service [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.132 162092 DEBUG oslo_service.service [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.133 162092 DEBUG oslo_service.service [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.133 162092 DEBUG oslo_service.service [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.133 162092 DEBUG oslo_service.service [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.133 162092 DEBUG oslo_service.service [-] max_dns_nameservers = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.133 162092 DEBUG oslo_service.service [-] max_header_line = 16384 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.134 162092 DEBUG oslo_service.service [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.134 162092 DEBUG oslo_service.service [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.134 162092 DEBUG oslo_service.service [-] max_subnet_host_routes = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.134 162092 DEBUG oslo_service.service [-] metadata_backlog = 4096 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.134 162092 DEBUG oslo_service.service [-] metadata_proxy_group = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.135 162092 DEBUG oslo_service.service [-] metadata_proxy_shared_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.135 162092 DEBUG oslo_service.service [-] metadata_proxy_socket = /var/lib/neutron/metadata_proxy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.135 162092 DEBUG oslo_service.service [-] metadata_proxy_socket_mode = deduce log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.135 162092 DEBUG oslo_service.service [-] metadata_proxy_user = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.136 162092 DEBUG oslo_service.service [-] metadata_workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.136 162092 DEBUG oslo_service.service [-] network_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.136 162092 DEBUG oslo_service.service [-] notify_nova_on_port_data_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.136 162092 DEBUG oslo_service.service [-] notify_nova_on_port_status_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.136 162092 DEBUG oslo_service.service [-] nova_client_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.137 162092 DEBUG oslo_service.service [-] nova_client_priv_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.137 162092 DEBUG oslo_service.service [-] nova_metadata_host = nova-metadata-internal.openstack.svc log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.137 162092 DEBUG oslo_service.service [-] nova_metadata_insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.137 162092 DEBUG oslo_service.service [-] nova_metadata_port = 8775 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.138 162092 DEBUG oslo_service.service [-] nova_metadata_protocol = http log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.138 162092 DEBUG oslo_service.service [-] pagination_max_limit = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.138 162092 DEBUG oslo_service.service [-] periodic_fuzzy_delay = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.138 162092 DEBUG oslo_service.service [-] periodic_interval = 40 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.139 162092 DEBUG oslo_service.service [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.139 162092 DEBUG oslo_service.service [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.139 162092 DEBUG oslo_service.service [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.139 162092 DEBUG oslo_service.service [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.139 162092 DEBUG oslo_service.service [-] retry_until_window = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.140 162092 DEBUG oslo_service.service [-] rpc_resources_processing_step = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.140 162092 DEBUG oslo_service.service [-] rpc_response_max_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.140 162092 DEBUG oslo_service.service [-] rpc_state_report_workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.140 162092 DEBUG oslo_service.service [-] rpc_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.140 162092 DEBUG oslo_service.service [-] send_events_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.141 162092 DEBUG oslo_service.service [-] service_plugins = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.141 162092 DEBUG oslo_service.service [-] setproctitle = on log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.141 162092 DEBUG oslo_service.service [-] state_path = /var/lib/neutron log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.141 162092 DEBUG oslo_service.service [-] syslog_log_facility = syslog log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.141 162092 DEBUG oslo_service.service [-] tcp_keepidle = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.142 162092 DEBUG oslo_service.service [-] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.142 162092 DEBUG oslo_service.service [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.142 162092 DEBUG oslo_service.service [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.142 162092 DEBUG oslo_service.service [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.143 162092 DEBUG oslo_service.service [-] use_ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.143 162092 DEBUG oslo_service.service [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.143 162092 DEBUG oslo_service.service [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.143 162092 DEBUG oslo_service.service [-] vlan_transparent = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.143 162092 DEBUG oslo_service.service [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.144 162092 DEBUG oslo_service.service [-] wsgi_default_pool_size = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.144 162092 DEBUG oslo_service.service [-] wsgi_keep_alive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.144 162092 DEBUG oslo_service.service [-] wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.144 162092 DEBUG oslo_service.service [-] wsgi_server_debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.144 162092 DEBUG oslo_service.service [-] oslo_concurrency.disable_process_locking = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.145 162092 DEBUG oslo_service.service [-] oslo_concurrency.lock_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.145 162092 DEBUG oslo_service.service [-] profiler.connection_string = messaging:// log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.145 162092 DEBUG oslo_service.service [-] profiler.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.145 162092 DEBUG oslo_service.service [-] profiler.es_doc_type = notification log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.146 162092 DEBUG oslo_service.service [-] profiler.es_scroll_size = 10000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.146 162092 DEBUG oslo_service.service [-] profiler.es_scroll_time = 2m log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.146 162092 DEBUG oslo_service.service [-] profiler.filter_error_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.146 162092 DEBUG oslo_service.service [-] profiler.hmac_keys = SECRET_KEY log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.147 162092 DEBUG oslo_service.service [-] profiler.sentinel_service_name = mymaster log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.147 162092 DEBUG oslo_service.service [-] profiler.socket_timeout = 0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.147 162092 DEBUG oslo_service.service [-] profiler.trace_sqlalchemy = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.147 162092 DEBUG oslo_service.service [-] oslo_policy.enforce_new_defaults = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.147 162092 DEBUG oslo_service.service [-] oslo_policy.enforce_scope = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.148 162092 DEBUG oslo_service.service [-] oslo_policy.policy_default_rule = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.148 162092 DEBUG oslo_service.service [-] oslo_policy.policy_dirs = ['policy.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.148 162092 DEBUG oslo_service.service [-] oslo_policy.policy_file = policy.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.148 162092 DEBUG oslo_service.service [-] oslo_policy.remote_content_type = application/x-www-form-urlencoded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.149 162092 DEBUG oslo_service.service [-] oslo_policy.remote_ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.149 162092 DEBUG oslo_service.service [-] oslo_policy.remote_ssl_client_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.149 162092 DEBUG oslo_service.service [-] oslo_policy.remote_ssl_client_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.149 162092 DEBUG oslo_service.service [-] oslo_policy.remote_ssl_verify_server_crt = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.150 162092 DEBUG oslo_service.service [-] oslo_messaging_metrics.metrics_buffer_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.150 162092 DEBUG oslo_service.service [-] oslo_messaging_metrics.metrics_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.150 162092 DEBUG oslo_service.service [-] oslo_messaging_metrics.metrics_process_name = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.150 162092 DEBUG oslo_service.service [-] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.150 162092 DEBUG oslo_service.service [-] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.151 162092 DEBUG oslo_service.service [-] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.151 162092 DEBUG oslo_service.service [-] service_providers.service_provider = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.151 162092 DEBUG oslo_service.service [-] privsep.capabilities = [21, 12, 1, 2, 19] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.151 162092 DEBUG oslo_service.service [-] privsep.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.152 162092 DEBUG oslo_service.service [-] privsep.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.152 162092 DEBUG oslo_service.service [-] privsep.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.152 162092 DEBUG oslo_service.service [-] privsep.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.152 162092 DEBUG oslo_service.service [-] privsep.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.152 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.capabilities = [21, 12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.153 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.153 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.153 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.153 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.153 162092 DEBUG oslo_service.service [-] privsep_dhcp_release.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.154 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.capabilities = [21, 12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.154 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.155 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.155 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.155 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.155 162092 DEBUG oslo_service.service [-] privsep_ovs_vsctl.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.155 162092 DEBUG oslo_service.service [-] privsep_namespace.capabilities = [21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.156 162092 DEBUG oslo_service.service [-] privsep_namespace.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.156 162092 DEBUG oslo_service.service [-] privsep_namespace.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.156 162092 DEBUG oslo_service.service [-] privsep_namespace.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.156 162092 DEBUG oslo_service.service [-] privsep_namespace.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.156 162092 DEBUG oslo_service.service [-] privsep_namespace.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.157 162092 DEBUG oslo_service.service [-] privsep_conntrack.capabilities = [12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.157 162092 DEBUG oslo_service.service [-] privsep_conntrack.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.157 162092 DEBUG oslo_service.service [-] privsep_conntrack.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.157 162092 DEBUG oslo_service.service [-] privsep_conntrack.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.157 162092 DEBUG oslo_service.service [-] privsep_conntrack.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.158 162092 DEBUG oslo_service.service [-] privsep_conntrack.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.158 162092 DEBUG oslo_service.service [-] privsep_link.capabilities = [12, 21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.158 162092 DEBUG oslo_service.service [-] privsep_link.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.158 162092 DEBUG oslo_service.service [-] privsep_link.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.158 162092 DEBUG oslo_service.service [-] privsep_link.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.159 162092 DEBUG oslo_service.service [-] privsep_link.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.159 162092 DEBUG oslo_service.service [-] privsep_link.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.159 162092 DEBUG oslo_service.service [-] AGENT.check_child_processes_action = respawn log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.159 162092 DEBUG oslo_service.service [-] AGENT.check_child_processes_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.160 162092 DEBUG oslo_service.service [-] AGENT.comment_iptables_rules = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.160 162092 DEBUG oslo_service.service [-] AGENT.debug_iptables_rules = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.160 162092 DEBUG oslo_service.service [-] AGENT.kill_scripts_path = /etc/neutron/kill_scripts/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.160 162092 DEBUG oslo_service.service [-] AGENT.root_helper = sudo neutron-rootwrap /etc/neutron/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.160 162092 DEBUG oslo_service.service [-] AGENT.root_helper_daemon = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.161 162092 DEBUG oslo_service.service [-] AGENT.use_helper_for_ns_read = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.161 162092 DEBUG oslo_service.service [-] AGENT.use_random_fully = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.161 162092 DEBUG oslo_service.service [-] oslo_versionedobjects.fatal_exception_format_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.161 162092 DEBUG oslo_service.service [-] QUOTAS.default_quota = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.162 162092 DEBUG oslo_service.service [-] QUOTAS.quota_driver = neutron.db.quota.driver_nolock.DbQuotaNoLockDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.162 162092 DEBUG oslo_service.service [-] QUOTAS.quota_network = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.162 162092 DEBUG oslo_service.service [-] QUOTAS.quota_port = 500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.162 162092 DEBUG oslo_service.service [-] QUOTAS.quota_security_group = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] QUOTAS.quota_security_group_rule = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] QUOTAS.quota_subnet = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] QUOTAS.track_quota_usage = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] nova.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] nova.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] nova.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.163 162092 DEBUG oslo_service.service [-] nova.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.endpoint_type = public log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.164 162092 DEBUG oslo_service.service [-] nova.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.endpoint_type = public log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.165 162092 DEBUG oslo_service.service [-] placement.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] placement.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] placement.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] placement.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] placement.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] ironic.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] ironic.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.166 162092 DEBUG oslo_service.service [-] ironic.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.enable_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.167 162092 DEBUG oslo_service.service [-] ironic.interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.service_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.168 162092 DEBUG oslo_service.service [-] ironic.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ironic.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ironic.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ironic.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ironic.valid_interfaces = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ironic.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] cli_script.dry_run = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ovn.allow_stateless_action_supported = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.169 162092 DEBUG oslo_service.service [-] ovn.dhcp_default_lease_time = 43200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.disable_ovn_dhcp_for_baremetal_ports = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.dns_servers = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.enable_distributed_floating_ip = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.neutron_sync_mode = log log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.ovn_dhcp4_global_options = {} log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.ovn_dhcp6_global_options = {} log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.170 162092 DEBUG oslo_service.service [-] ovn.ovn_emit_need_to_frag = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_l3_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_l3_scheduler = leastloaded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_metadata_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_nb_ca_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_nb_certificate = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_nb_connection = tcp:127.0.0.1:6641 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.171 162092 DEBUG oslo_service.service [-] ovn.ovn_nb_private_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovn_sb_ca_cert = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovn_sb_certificate = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovn_sb_connection = tcp:ovsdbserver-sb.openstack.svc:6642 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovn_sb_private_key = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovsdb_connection_timeout = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovsdb_log_level = INFO log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.172 162092 DEBUG oslo_service.service [-] ovn.ovsdb_probe_interval = 60000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] ovn.ovsdb_retry_max_interval = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] ovn.vhost_sock_dir = /var/run/openvswitch log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] ovn.vif_type = ovs log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] OVS.bridge_mac_table_size = 50000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] OVS.igmp_snooping_enable = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] OVS.ovsdb_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.173 162092 DEBUG oslo_service.service [-] ovs.ovsdb_connection = tcp:127.0.0.1:6640 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] ovs.ovsdb_connection_timeout = 180 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.amqp_durable_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.174 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.175 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.176 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_quorum_queue = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.177 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_notifications.driver = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.178 162092 DEBUG oslo_service.service [-] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.179 162092 DEBUG oslo_service.service [-] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:25:48 localhost ovn_metadata_agent[162087]: 2025-11-27 09:25:48.179 162092 DEBUG oslo_service.service [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613#033[00m Nov 27 04:25:51 localhost sshd[162232]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:25:51 localhost systemd-logind[761]: New session 53 of user zuul. Nov 27 04:25:51 localhost systemd[1]: Started Session 53 of User zuul. Nov 27 04:25:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18444 DF PROTO=TCP SPT=52324 DPT=9100 SEQ=1760260511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C809B0000000001030307) Nov 27 04:25:52 localhost python3.9[162325]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:25:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36924 DF PROTO=TCP SPT=38924 DPT=9101 SEQ=4068888865 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C889A0000000001030307) Nov 27 04:25:53 localhost python3.9[162421]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps -a --filter name=^nova_virtlogd$ --format \{\{.Names\}\} _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:25:54 localhost python3.9[162527]: ansible-ansible.legacy.command Invoked with _raw_params=podman stop nova_virtlogd _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:25:54 localhost systemd[1]: libpod-6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e.scope: Deactivated successfully. Nov 27 04:25:54 localhost podman[162528]: 2025-11-27 09:25:54.82399594 +0000 UTC m=+0.075424491 container died 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, name=rhosp17/openstack-nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-type=git, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.buildah.version=1.41.4, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, release=1761123044, tcib_managed=true, vendor=Red Hat, Inc., distribution-scope=public, konflux.additional-tags=17.1.12 17.1_20251118.1, build-date=2025-11-19T00:35:22Z, io.openshift.expose-services=, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d) Nov 27 04:25:54 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e-userdata-shm.mount: Deactivated successfully. Nov 27 04:25:54 localhost podman[162528]: 2025-11-27 09:25:54.861217412 +0000 UTC m=+0.112645913 container cleanup 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, vendor=Red Hat, Inc., baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, tcib_managed=true, konflux.additional-tags=17.1.12 17.1_20251118.1, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, url=https://www.redhat.com, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, description=Red Hat OpenStack Platform 17.1 nova-libvirt, com.redhat.component=openstack-nova-libvirt-container, io.openshift.expose-services=, distribution-scope=public, io.buildah.version=1.41.4, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, version=17.1.12, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, release=1761123044, vcs-type=git, batch=17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-libvirt, build-date=2025-11-19T00:35:22Z) Nov 27 04:25:54 localhost podman[162541]: 2025-11-27 09:25:54.921936992 +0000 UTC m=+0.085712910 container remove 6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e (image=registry.redhat.io/rhosp-rhel9/openstack-nova-libvirt:17.1, name=nova_virtlogd, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-libvirt, release=1761123044, vcs-type=git, version=17.1.12, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.component=openstack-nova-libvirt-container, io.buildah.version=1.41.4, description=Red Hat OpenStack Platform 17.1 nova-libvirt, architecture=x86_64, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, maintainer=OpenStack TripleO Team, name=rhosp17/openstack-nova-libvirt, url=https://www.redhat.com, summary=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-libvirt, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, batch=17.1_20251118.1, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-libvirt, io.openshift.expose-services=, tcib_managed=true, distribution-scope=public, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-11-19T00:35:22Z) Nov 27 04:25:54 localhost systemd[1]: libpod-conmon-6ab87355f4e0f4259af254d83ee60e371e315794f41998d349c56da64113883e.scope: Deactivated successfully. Nov 27 04:25:55 localhost systemd[1]: var-lib-containers-storage-overlay-d0f444e3451054fa27276f38df962bce4740a4eeaa689d937364bf82d226519f-merged.mount: Deactivated successfully. Nov 27 04:25:56 localhost python3.9[162650]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:25:56 localhost systemd[1]: Reloading. Nov 27 04:25:56 localhost systemd-rc-local-generator[162671]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:25:56 localhost systemd-sysv-generator[162680]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:25:56 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53178 DF PROTO=TCP SPT=34326 DPT=9101 SEQ=1868278541 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C93DB0000000001030307) Nov 27 04:25:57 localhost python3.9[162776]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:25:57 localhost network[162793]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:25:57 localhost network[162794]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:25:57 localhost network[162795]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:25:59 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:25:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=3459 DF PROTO=TCP SPT=60552 DPT=9105 SEQ=1626599328 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128C9FDA0000000001030307) Nov 27 04:26:02 localhost python3.9[162996]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_libvirt.target state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18384 DF PROTO=TCP SPT=59718 DPT=9105 SEQ=699390108 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CAC5A0000000001030307) Nov 27 04:26:02 localhost systemd[1]: Reloading. Nov 27 04:26:03 localhost systemd-rc-local-generator[163024]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:26:03 localhost systemd-sysv-generator[163028]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:26:03 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:26:03 localhost systemd[1]: Stopped target tripleo_nova_libvirt.target. Nov 27 04:26:04 localhost python3.9[163128]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtlogd_wrapper.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:04 localhost python3.9[163221]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtnodedevd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:05 localhost python3.9[163314]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtproxyd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:06 localhost python3.9[163422]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtqemud.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:26:06 localhost podman[163441]: 2025-11-27 09:26:06.728707143 +0000 UTC m=+0.073698134 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_controller, tcib_managed=true) Nov 27 04:26:06 localhost podman[163441]: 2025-11-27 09:26:06.759909871 +0000 UTC m=+0.104900832 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.vendor=CentOS, config_id=ovn_controller, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:26:06 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:26:07 localhost python3.9[163575]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtsecretd.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36927 DF PROTO=TCP SPT=38924 DPT=9101 SEQ=4068888865 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CBFDA0000000001030307) Nov 27 04:26:07 localhost python3.9[163678]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_virtstoraged.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:26:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64022 DF PROTO=TCP SPT=50710 DPT=9882 SEQ=1712923357 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CC3470000000001030307) Nov 27 04:26:11 localhost python3.9[163786]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_libvirt.target state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:11 localhost python3.9[163878]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtlogd_wrapper.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64024 DF PROTO=TCP SPT=50710 DPT=9882 SEQ=1712923357 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CCF5A0000000001030307) Nov 27 04:26:12 localhost python3.9[163970]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtnodedevd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:26:12 localhost systemd[1]: tmp-crun.k3H5g7.mount: Deactivated successfully. Nov 27 04:26:12 localhost podman[164063]: 2025-11-27 09:26:12.829513374 +0000 UTC m=+0.097552583 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_managed=true, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 04:26:12 localhost podman[164063]: 2025-11-27 09:26:12.864009032 +0000 UTC m=+0.132048221 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent) Nov 27 04:26:12 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:26:12 localhost python3.9[164062]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtproxyd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:13 localhost python3.9[164174]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtqemud.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:14 localhost python3.9[164266]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtsecretd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:15 localhost python3.9[164358]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_virtstoraged.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40095 DF PROTO=TCP SPT=46256 DPT=9102 SEQ=311731591 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CDDD90000000001030307) Nov 27 04:26:16 localhost python3.9[164450]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_libvirt.target state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:17 localhost python3.9[164542]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtlogd_wrapper.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:17 localhost python3.9[164634]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtnodedevd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49545 DF PROTO=TCP SPT=58812 DPT=9100 SEQ=554947251 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CE61A0000000001030307) Nov 27 04:26:18 localhost python3.9[164726]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtproxyd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:18 localhost python3.9[164818]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtqemud.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:19 localhost python3.9[164910]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtsecretd.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:20 localhost python3.9[165002]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_virtstoraged.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:26:20 localhost python3.9[165094]: ansible-ansible.legacy.command Invoked with _raw_params=if systemctl is-active certmonger.service; then#012 systemctl disable --now certmonger.service#012 test -f /etc/systemd/system/certmonger.service || systemctl mask certmonger.service#012fi#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49546 DF PROTO=TCP SPT=58812 DPT=9100 SEQ=554947251 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CF5DA0000000001030307) Nov 27 04:26:21 localhost python3.9[165186]: ansible-ansible.builtin.find Invoked with file_type=any hidden=True paths=['/var/lib/certmonger/requests'] patterns=[] read_whole_file=False age_stamp=mtime recurse=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:26:23 localhost python3.9[165278]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:26:23 localhost systemd[1]: Reloading. Nov 27 04:26:23 localhost systemd-sysv-generator[165304]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:26:23 localhost systemd-rc-local-generator[165300]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:26:23 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:26:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45790 DF PROTO=TCP SPT=44500 DPT=9101 SEQ=3750612685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128CFD9B0000000001030307) Nov 27 04:26:24 localhost python3.9[165405]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_libvirt.target _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:24 localhost python3.9[165498]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtlogd_wrapper.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:25 localhost python3.9[165591]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtnodedevd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:25 localhost python3.9[165684]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtproxyd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13164 DF PROTO=TCP SPT=40708 DPT=9105 SEQ=428360410 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D09DB0000000001030307) Nov 27 04:26:27 localhost python3.9[165777]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtqemud.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:28 localhost python3.9[165870]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtsecretd.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:29 localhost python3.9[165963]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_virtstoraged.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:26:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45792 DF PROTO=TCP SPT=44500 DPT=9101 SEQ=3750612685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D155A0000000001030307) Nov 27 04:26:30 localhost python3.9[166056]: ansible-ansible.builtin.getent Invoked with database=passwd key=libvirt fail_key=True service=None split=None Nov 27 04:26:31 localhost python3.9[166149]: ansible-ansible.builtin.group Invoked with gid=42473 name=libvirt state=present force=False system=False local=False non_unique=False gid_min=None gid_max=None Nov 27 04:26:32 localhost python3.9[166247]: ansible-ansible.builtin.user Invoked with comment=libvirt user group=libvirt groups=[''] name=libvirt shell=/sbin/nologin state=present uid=42473 non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.localdomain update_password=always home=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None password_expire_account_disable=None uid_min=None uid_max=None Nov 27 04:26:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13166 DF PROTO=TCP SPT=40708 DPT=9105 SEQ=428360410 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D219A0000000001030307) Nov 27 04:26:33 localhost python3.9[166347]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:26:34 localhost python3.9[166401]: ansible-ansible.legacy.dnf Invoked with name=['libvirt ', 'libvirt-admin ', 'libvirt-client ', 'libvirt-daemon ', 'qemu-kvm', 'qemu-img', 'libguestfs', 'libseccomp', 'swtpm', 'swtpm-tools', 'edk2-ovmf', 'ceph-common', 'cyrus-sasl-scram'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:26:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:26:36 localhost podman[166404]: 2025-11-27 09:26:36.982181641 +0000 UTC m=+0.077782559 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:26:37 localhost podman[166404]: 2025-11-27 09:26:37.050689432 +0000 UTC m=+0.146290390 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller) Nov 27 04:26:37 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:26:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45793 DF PROTO=TCP SPT=44500 DPT=9101 SEQ=3750612685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D35DA0000000001030307) Nov 27 04:26:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46443 DF PROTO=TCP SPT=48136 DPT=9882 SEQ=3516197585 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D38760000000001030307) Nov 27 04:26:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46445 DF PROTO=TCP SPT=48136 DPT=9882 SEQ=3516197585 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D449A0000000001030307) Nov 27 04:26:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:26:42 localhost podman[166497]: 2025-11-27 09:26:42.975274786 +0000 UTC m=+0.073017323 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible) Nov 27 04:26:43 localhost podman[166497]: 2025-11-27 09:26:43.012789673 +0000 UTC m=+0.110532250 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:26:43 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:26:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:26:43.532 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:26:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:26:43.532 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:26:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:26:43.534 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:26:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=64656 DF PROTO=TCP SPT=57676 DPT=9102 SEQ=416822806 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D530A0000000001030307) Nov 27 04:26:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42858 DF PROTO=TCP SPT=57622 DPT=9100 SEQ=2115111719 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D5B1A0000000001030307) Nov 27 04:26:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42859 DF PROTO=TCP SPT=57622 DPT=9100 SEQ=2115111719 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D6ADB0000000001030307) Nov 27 04:26:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21812 DF PROTO=TCP SPT=45956 DPT=9101 SEQ=1847216372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D72DB0000000001030307) Nov 27 04:26:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=50224 DF PROTO=TCP SPT=54656 DPT=9105 SEQ=354621010 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D7F1A0000000001030307) Nov 27 04:26:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18387 DF PROTO=TCP SPT=59718 DPT=9105 SEQ=699390108 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D89DB0000000001030307) Nov 27 04:27:00 localhost kernel: SELinux: Converting 2760 SID table entries... Nov 27 04:27:00 localhost kernel: SELinux: Context system_u:object_r:insights_client_cache_t:s0 became invalid (unmapped). Nov 27 04:27:00 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:00 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:00 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:00 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:00 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:00 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:00 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=50226 DF PROTO=TCP SPT=54656 DPT=9105 SEQ=354621010 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128D96DB0000000001030307) Nov 27 04:27:07 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=19 res=1 Nov 27 04:27:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:27:08 localhost systemd[1]: tmp-crun.8lDyY5.mount: Deactivated successfully. Nov 27 04:27:08 localhost podman[167580]: 2025-11-27 09:27:08.026679634 +0000 UTC m=+0.108237229 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:27:08 localhost podman[167580]: 2025-11-27 09:27:08.113861216 +0000 UTC m=+0.195418801 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=ovn_controller, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:27:08 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:27:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=21815 DF PROTO=TCP SPT=45956 DPT=9101 SEQ=1847216372 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DABDA0000000001030307) Nov 27 04:27:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24132 DF PROTO=TCP SPT=46764 DPT=9882 SEQ=2415936180 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DADA70000000001030307) Nov 27 04:27:10 localhost kernel: SELinux: Converting 2763 SID table entries... Nov 27 04:27:10 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:10 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:10 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:10 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:10 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:10 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:10 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24134 DF PROTO=TCP SPT=46764 DPT=9882 SEQ=2415936180 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DB99A0000000001030307) Nov 27 04:27:13 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=20 res=1 Nov 27 04:27:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:27:13 localhost podman[167681]: 2025-11-27 09:27:13.998634709 +0000 UTC m=+0.087464900 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS) Nov 27 04:27:14 localhost podman[167681]: 2025-11-27 09:27:14.033881077 +0000 UTC m=+0.122711258 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true) Nov 27 04:27:14 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:27:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15645 DF PROTO=TCP SPT=40728 DPT=9102 SEQ=1740913125 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DC8390000000001030307) Nov 27 04:27:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12560 DF PROTO=TCP SPT=34908 DPT=9100 SEQ=3543078961 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DD05B0000000001030307) Nov 27 04:27:19 localhost kernel: SELinux: Converting 2763 SID table entries... Nov 27 04:27:19 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:19 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:19 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:19 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:19 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:19 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:19 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12561 DF PROTO=TCP SPT=34908 DPT=9100 SEQ=3543078961 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DE01A0000000001030307) Nov 27 04:27:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38887 DF PROTO=TCP SPT=60906 DPT=9101 SEQ=2671565859 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DE81A0000000001030307) Nov 27 04:27:26 localhost sshd[167706]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:27:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45795 DF PROTO=TCP SPT=44500 DPT=9101 SEQ=3750612685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DF3DA0000000001030307) Nov 27 04:27:27 localhost kernel: SELinux: Converting 2763 SID table entries... Nov 27 04:27:27 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:27 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:27 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:27 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:27 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:27 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:27 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38889 DF PROTO=TCP SPT=60906 DPT=9101 SEQ=2671565859 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128DFFDA0000000001030307) Nov 27 04:27:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7848 DF PROTO=TCP SPT=47468 DPT=9105 SEQ=1747567278 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E0BDA0000000001030307) Nov 27 04:27:37 localhost kernel: SELinux: Converting 2763 SID table entries... Nov 27 04:27:37 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:37 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:37 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:37 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:37 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:37 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:37 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38890 DF PROTO=TCP SPT=60906 DPT=9101 SEQ=2671565859 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E1FDB0000000001030307) Nov 27 04:27:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13427 DF PROTO=TCP SPT=34768 DPT=9882 SEQ=2093068775 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E22D60000000001030307) Nov 27 04:27:38 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=23 res=1 Nov 27 04:27:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:27:38 localhost podman[167727]: 2025-11-27 09:27:38.991109151 +0000 UTC m=+0.080644350 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller) Nov 27 04:27:39 localhost podman[167727]: 2025-11-27 09:27:39.033880642 +0000 UTC m=+0.123415841 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:27:39 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:27:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13429 DF PROTO=TCP SPT=34768 DPT=9882 SEQ=2093068775 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E2EDB0000000001030307) Nov 27 04:27:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:27:43.533 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:27:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:27:43.533 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:27:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:27:43.535 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:27:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:27:45 localhost podman[167752]: 2025-11-27 09:27:45.002633828 +0000 UTC m=+0.102563539 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:27:45 localhost podman[167752]: 2025-11-27 09:27:45.038077632 +0000 UTC m=+0.138007383 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:27:45 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:27:45 localhost kernel: SELinux: Converting 2763 SID table entries... Nov 27 04:27:45 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:45 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:45 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:45 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:45 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:45 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:45 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59244 DF PROTO=TCP SPT=44346 DPT=9102 SEQ=2517940444 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E3D6B0000000001030307) Nov 27 04:27:45 localhost systemd[1]: Reloading. Nov 27 04:27:45 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=24 res=1 Nov 27 04:27:45 localhost systemd-rc-local-generator[167800]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:27:45 localhost systemd-sysv-generator[167805]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:27:45 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:27:46 localhost systemd[1]: Reloading. Nov 27 04:27:46 localhost systemd-rc-local-generator[167837]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:27:46 localhost systemd-sysv-generator[167843]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:27:46 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:27:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49964 DF PROTO=TCP SPT=41874 DPT=9100 SEQ=2051233298 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E459A0000000001030307) Nov 27 04:27:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:27:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 6000.1 total, 600.0 interval#012Cumulative writes: 5014 writes, 22K keys, 5014 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5014 writes, 579 syncs, 8.66 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.1 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 11 last_copies: 8 last_secs: 3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.1 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56526f65b610#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 11 last_copies: 8 last_secs: 3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.1 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdo Nov 27 04:27:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49965 DF PROTO=TCP SPT=41874 DPT=9100 SEQ=2051233298 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E555A0000000001030307) Nov 27 04:27:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23334 DF PROTO=TCP SPT=38572 DPT=9101 SEQ=873108632 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E5D5A0000000001030307) Nov 27 04:27:55 localhost kernel: SELinux: Converting 2764 SID table entries... Nov 27 04:27:55 localhost kernel: SELinux: policy capability network_peer_controls=1 Nov 27 04:27:55 localhost kernel: SELinux: policy capability open_perms=1 Nov 27 04:27:55 localhost kernel: SELinux: policy capability extended_socket_class=1 Nov 27 04:27:55 localhost kernel: SELinux: policy capability always_check_network=0 Nov 27 04:27:55 localhost kernel: SELinux: policy capability cgroup_seclabel=1 Nov 27 04:27:55 localhost kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 27 04:27:55 localhost kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Nov 27 04:27:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:27:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 6000.2 total, 600.0 interval#012Cumulative writes: 5569 writes, 24K keys, 5569 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5569 writes, 741 syncs, 7.52 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 2/0 2.61 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Sum 2/0 2.61 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.1 0.02 0.00 1 0.017 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.2 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 11 last_copies: 8 last_secs: 2.8e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] **#012#012** Compaction Stats [m-0] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-0] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.2 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x5568f618c2d0#2 capacity: 1.62 GB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 11 last_copies: 8 last_secs: 2.8e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,8.34465e-05%) FilterBlock(3,0.33 KB,1.92569e-05%) IndexBlock(3,0.34 KB,2.01739e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [m-0] **#012#012** Compaction Stats [m-1] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0#012#012** Compaction Stats [m-1] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 6000.2 total, 4800.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_sl Nov 27 04:27:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42466 DF PROTO=TCP SPT=38136 DPT=9105 SEQ=1276197881 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E695B0000000001030307) Nov 27 04:27:59 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 04:27:59 localhost dbus-broker-launch[756]: avc: op=load_policy lsm=selinux seqno=25 res=1 Nov 27 04:27:59 localhost dbus-broker-launch[752]: Noticed file-system modification, trigger reload. Nov 27 04:27:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23336 DF PROTO=TCP SPT=38572 DPT=9101 SEQ=873108632 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E751A0000000001030307) Nov 27 04:28:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42468 DF PROTO=TCP SPT=38136 DPT=9105 SEQ=1276197881 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E811A0000000001030307) Nov 27 04:28:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23337 DF PROTO=TCP SPT=38572 DPT=9101 SEQ=873108632 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E95DA0000000001030307) Nov 27 04:28:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55747 DF PROTO=TCP SPT=54434 DPT=9882 SEQ=1363585519 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128E98060000000001030307) Nov 27 04:28:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:28:09 localhost podman[168113]: 2025-11-27 09:28:09.671708656 +0000 UTC m=+0.117885622 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, config_id=ovn_controller, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 04:28:09 localhost podman[168113]: 2025-11-27 09:28:09.709041361 +0000 UTC m=+0.155218347 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:28:09 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:28:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55749 DF PROTO=TCP SPT=54434 DPT=9882 SEQ=1363585519 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EA41B0000000001030307) Nov 27 04:28:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13060 DF PROTO=TCP SPT=52400 DPT=9102 SEQ=3680730271 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EB29A0000000001030307) Nov 27 04:28:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:28:16 localhost podman[170542]: 2025-11-27 09:28:16.006817099 +0000 UTC m=+0.083142058 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 04:28:16 localhost podman[170542]: 2025-11-27 09:28:16.037585557 +0000 UTC m=+0.113910546 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:28:16 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:28:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47249 DF PROTO=TCP SPT=58358 DPT=9100 SEQ=834308005 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EBADA0000000001030307) Nov 27 04:28:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47250 DF PROTO=TCP SPT=58358 DPT=9100 SEQ=834308005 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128ECA9A0000000001030307) Nov 27 04:28:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42139 DF PROTO=TCP SPT=47002 DPT=9101 SEQ=3188616053 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128ED25A0000000001030307) Nov 27 04:28:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38892 DF PROTO=TCP SPT=60906 DPT=9101 SEQ=2671565859 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EDDDA0000000001030307) Nov 27 04:28:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=7851 DF PROTO=TCP SPT=47468 DPT=9105 SEQ=1747567278 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EE9DB0000000001030307) Nov 27 04:28:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15001 DF PROTO=TCP SPT=39520 DPT=9105 SEQ=2212999635 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128EF65A0000000001030307) Nov 27 04:28:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42142 DF PROTO=TCP SPT=47002 DPT=9101 SEQ=3188616053 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F09DA0000000001030307) Nov 27 04:28:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49069 DF PROTO=TCP SPT=50450 DPT=9882 SEQ=3293951119 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F0D370000000001030307) Nov 27 04:28:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:28:39 localhost podman[185122]: 2025-11-27 09:28:39.97406224 +0000 UTC m=+0.063384868 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 04:28:40 localhost podman[185122]: 2025-11-27 09:28:40.076894912 +0000 UTC m=+0.166217550 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:28:40 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:28:41 localhost systemd[1]: Stopping OpenSSH server daemon... Nov 27 04:28:41 localhost systemd[1]: sshd.service: Deactivated successfully. Nov 27 04:28:41 localhost systemd[1]: Stopped OpenSSH server daemon. Nov 27 04:28:41 localhost systemd[1]: sshd.service: Consumed 1.001s CPU time, read 32.0K from disk, written 0B to disk. Nov 27 04:28:41 localhost systemd[1]: Stopped target sshd-keygen.target. Nov 27 04:28:41 localhost systemd[1]: Stopping sshd-keygen.target... Nov 27 04:28:41 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:28:41 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:28:41 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). Nov 27 04:28:41 localhost systemd[1]: Reached target sshd-keygen.target. Nov 27 04:28:41 localhost systemd[1]: Starting OpenSSH server daemon... Nov 27 04:28:41 localhost sshd[185808]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:28:41 localhost systemd[1]: Started OpenSSH server daemon. Nov 27 04:28:41 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:41 localhost systemd[1]: /usr/lib/systemd/system/libvirtd.service:29: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:41 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:41 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:41 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=49071 DF PROTO=TCP SPT=50450 DPT=9882 SEQ=3293951119 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F195B0000000001030307) Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:42 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:28:43.533 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:28:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:28:43.534 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:28:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:28:43.535 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:28:43 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 04:28:43 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 04:28:43 localhost systemd[1]: Reloading. Nov 27 04:28:43 localhost systemd-rc-local-generator[186040]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:43 localhost systemd-sysv-generator[186044]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/libvirtd.service:29: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:44 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 04:28:44 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 04:28:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=61760 DF PROTO=TCP SPT=42850 DPT=9102 SEQ=3150171809 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F27C90000000001030307) Nov 27 04:28:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:28:46 localhost podman[189278]: 2025-11-27 09:28:46.507543703 +0000 UTC m=+0.097275769 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent) Nov 27 04:28:46 localhost podman[189278]: 2025-11-27 09:28:46.542911489 +0000 UTC m=+0.132643515 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, managed_by=edpm_ansible, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 04:28:46 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:28:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36434 DF PROTO=TCP SPT=49998 DPT=9100 SEQ=2510913687 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F2FDB0000000001030307) Nov 27 04:28:48 localhost python3.9[191307]: ansible-ansible.builtin.systemd Invoked with enabled=False masked=True name=libvirtd state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:28:48 localhost systemd[1]: Reloading. Nov 27 04:28:48 localhost systemd-sysv-generator[191753]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:48 localhost systemd-rc-local-generator[191750]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:48 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:48 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:49 localhost python3.9[191999]: ansible-ansible.builtin.systemd Invoked with enabled=False masked=True name=libvirtd-tcp.socket state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:28:49 localhost systemd[1]: Reloading. Nov 27 04:28:50 localhost systemd-rc-local-generator[192227]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:50 localhost systemd-sysv-generator[192231]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:50 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost python3.9[192842]: ansible-ansible.builtin.systemd Invoked with enabled=False masked=True name=libvirtd-tls.socket state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:28:51 localhost systemd[1]: Reloading. Nov 27 04:28:51 localhost systemd-rc-local-generator[193077]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:51 localhost systemd-sysv-generator[193082]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36435 DF PROTO=TCP SPT=49998 DPT=9100 SEQ=2510913687 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F3F9A0000000001030307) Nov 27 04:28:52 localhost python3.9[193478]: ansible-ansible.builtin.systemd Invoked with enabled=False masked=True name=virtproxyd-tcp.socket state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:28:52 localhost systemd[1]: Reloading. Nov 27 04:28:52 localhost systemd-sysv-generator[193727]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:52 localhost systemd-rc-local-generator[193722]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:52 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42392 DF PROTO=TCP SPT=39802 DPT=9101 SEQ=3226966282 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F479A0000000001030307) Nov 27 04:28:54 localhost python3.9[194762]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtlogd.service daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:28:55 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 04:28:55 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 04:28:55 localhost systemd[1]: man-db-cache-update.service: Consumed 14.228s CPU time. Nov 27 04:28:55 localhost systemd[1]: run-r46a31e1877494c31ba4cc15dc4f2efdb.service: Deactivated successfully. Nov 27 04:28:55 localhost systemd[1]: run-r6b38eef1b7b545339f70faf4e53a3a83.service: Deactivated successfully. Nov 27 04:28:55 localhost systemd[1]: Reloading. Nov 27 04:28:56 localhost systemd-rc-local-generator[195254]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:56 localhost systemd-sysv-generator[195259]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6038 DF PROTO=TCP SPT=34358 DPT=9105 SEQ=2862105774 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F53DA0000000001030307) Nov 27 04:28:57 localhost python3.9[195378]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtnodedevd.service daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:28:57 localhost systemd[1]: Reloading. Nov 27 04:28:57 localhost systemd-sysv-generator[195410]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:57 localhost systemd-rc-local-generator[195404]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:57 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost python3.9[195527]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtproxyd.service daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:28:58 localhost systemd[1]: Reloading. Nov 27 04:28:58 localhost systemd-sysv-generator[195560]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:28:58 localhost systemd-rc-local-generator[195555]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:58 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:28:59 localhost python3.9[195676]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtqemud.service daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:28:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42394 DF PROTO=TCP SPT=39802 DPT=9101 SEQ=3226966282 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F5F5A0000000001030307) Nov 27 04:29:00 localhost python3.9[195789]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtsecretd.service daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:00 localhost systemd[1]: Reloading. Nov 27 04:29:00 localhost systemd-rc-local-generator[195817]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:29:00 localhost systemd-sysv-generator[195821]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:00 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost python3.9[195937]: ansible-ansible.builtin.systemd Invoked with enabled=False masked=True name=virtproxyd-tls.socket state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None Nov 27 04:29:01 localhost systemd[1]: Reloading. Nov 27 04:29:01 localhost systemd-rc-local-generator[195964]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:29:01 localhost systemd-sysv-generator[195967]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:01 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:29:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6040 DF PROTO=TCP SPT=34358 DPT=9105 SEQ=2862105774 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F6B9A0000000001030307) Nov 27 04:29:03 localhost python3.9[196086]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtlogd.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:05 localhost python3.9[196199]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtlogd-admin.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:06 localhost python3.9[196312]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtnodedevd.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:07 localhost python3.9[196425]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtnodedevd-ro.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:07 localhost python3.9[196538]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtnodedevd-admin.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42395 DF PROTO=TCP SPT=39802 DPT=9101 SEQ=3226966282 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F7FDA0000000001030307) Nov 27 04:29:08 localhost python3.9[196651]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtproxyd.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38168 DF PROTO=TCP SPT=33088 DPT=9882 SEQ=4245649660 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F82670000000001030307) Nov 27 04:29:09 localhost python3.9[196764]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtproxyd-ro.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:10 localhost python3.9[196877]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtproxyd-admin.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:29:10 localhost podman[196879]: 2025-11-27 09:29:10.232896899 +0000 UTC m=+0.080978449 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:29:10 localhost podman[196879]: 2025-11-27 09:29:10.33902793 +0000 UTC m=+0.187109460 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_id=ovn_controller, container_name=ovn_controller) Nov 27 04:29:10 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:29:10 localhost python3.9[197014]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtqemud.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38170 DF PROTO=TCP SPT=33088 DPT=9882 SEQ=4245649660 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F8E5A0000000001030307) Nov 27 04:29:12 localhost python3.9[197195]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtqemud-ro.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:13 localhost python3.9[197326]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtqemud-admin.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:14 localhost python3.9[197439]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtsecretd.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:15 localhost python3.9[197552]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtsecretd-ro.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25474 DF PROTO=TCP SPT=39904 DPT=9102 SEQ=3233137267 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128F9CF90000000001030307) Nov 27 04:29:16 localhost python3.9[197665]: ansible-ansible.builtin.systemd Invoked with enabled=True masked=False name=virtsecretd-admin.socket daemon_reload=False daemon_reexec=False scope=system no_block=False state=None force=None Nov 27 04:29:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:29:17 localhost podman[197667]: 2025-11-27 09:29:17.003988209 +0000 UTC m=+0.095727692 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 04:29:17 localhost podman[197667]: 2025-11-27 09:29:17.009839989 +0000 UTC m=+0.101579502 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 04:29:17 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:29:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4474 DF PROTO=TCP SPT=50304 DPT=9100 SEQ=3120665478 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FA51A0000000001030307) Nov 27 04:29:21 localhost python3.9[197798]: ansible-ansible.builtin.file Invoked with group=root owner=root path=/etc/tmpfiles.d/ setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4475 DF PROTO=TCP SPT=50304 DPT=9100 SEQ=3120665478 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FB4DB0000000001030307) Nov 27 04:29:22 localhost python3.9[197908]: ansible-ansible.builtin.file Invoked with group=root owner=root path=/var/lib/edpm-config/firewall setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:22 localhost python3.9[198018]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/pki/libvirt setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:23 localhost python3.9[198128]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/pki/libvirt/private setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20717 DF PROTO=TCP SPT=52974 DPT=9101 SEQ=3419457726 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FBCDA0000000001030307) Nov 27 04:29:24 localhost python3.9[198238]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/pki/CA setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:24 localhost python3.9[198348]: ansible-ansible.builtin.file Invoked with group=qemu owner=root path=/etc/pki/qemu setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:29:25 localhost python3.9[198458]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/virtlogd.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:26 localhost python3.9[198548]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/virtlogd.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235764.9608624-1646-69349265109040/.source.conf follow=False _original_basename=virtlogd.conf checksum=d7a72ae92c2c205983b029473e05a6aa4c58ec24 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42231 DF PROTO=TCP SPT=34122 DPT=9105 SEQ=273137152 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FC8DB0000000001030307) Nov 27 04:29:26 localhost python3.9[198658]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/virtnodedevd.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:28 localhost python3.9[198748]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/virtnodedevd.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235766.5847502-1646-266444918878016/.source.conf follow=False _original_basename=virtnodedevd.conf checksum=7a604468adb2868f1ab6ebd0fd4622286e6373e2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:28 localhost python3.9[198858]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/virtproxyd.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:29 localhost python3.9[198948]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/virtproxyd.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235768.5064101-1646-230115675370933/.source.conf follow=False _original_basename=virtproxyd.conf checksum=28bc484b7c9988e03de49d4fcc0a088ea975f716 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15004 DF PROTO=TCP SPT=39520 DPT=9105 SEQ=2212999635 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FD3DA0000000001030307) Nov 27 04:29:30 localhost python3.9[199058]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/virtqemud.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:31 localhost python3.9[199148]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/virtqemud.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235770.27121-1646-276254113144489/.source.conf follow=False _original_basename=virtqemud.conf checksum=7a604468adb2868f1ab6ebd0fd4622286e6373e2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:31 localhost python3.9[199258]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/qemu.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:32 localhost python3.9[199348]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/qemu.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235771.3827908-1646-137783158874282/.source.conf follow=False _original_basename=qemu.conf.j2 checksum=8d9b2057482987a531d808ceb2ac4bc7d43bf17c backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42233 DF PROTO=TCP SPT=34122 DPT=9105 SEQ=273137152 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FE09A0000000001030307) Nov 27 04:29:33 localhost python3.9[199458]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/virtsecretd.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:33 localhost python3.9[199548]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/virtsecretd.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235772.5476704-1646-217080077493985/.source.conf follow=False _original_basename=virtsecretd.conf checksum=7a604468adb2868f1ab6ebd0fd4622286e6373e2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:34 localhost python3.9[199658]: ansible-ansible.legacy.stat Invoked with path=/etc/libvirt/auth.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:34 localhost python3.9[199746]: ansible-ansible.legacy.copy Invoked with dest=/etc/libvirt/auth.conf group=libvirt mode=0600 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235773.7442896-1646-66377759562875/.source.conf follow=False _original_basename=auth.conf checksum=da39a3ee5e6b4b0d3255bfef95601890afd80709 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:35 localhost python3.9[199856]: ansible-ansible.legacy.stat Invoked with path=/etc/sasl2/libvirt.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:36 localhost python3.9[199946]: ansible-ansible.legacy.copy Invoked with dest=/etc/sasl2/libvirt.conf group=libvirt mode=0640 owner=libvirt src=/home/zuul/.ansible/tmp/ansible-tmp-1764235774.8929515-1646-29994426857107/.source.conf follow=False _original_basename=sasl_libvirt.conf checksum=652e4d404bf79253d06956b8e9847c9364979d4a backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:36 localhost python3.9[200056]: ansible-ansible.builtin.file Invoked with path=/etc/libvirt/passwd.db state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:37 localhost python3.9[200166]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtlogd.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:38 localhost python3.9[200276]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtlogd-admin.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20720 DF PROTO=TCP SPT=52974 DPT=9101 SEQ=3419457726 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FF5DA0000000001030307) Nov 27 04:29:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39961 DF PROTO=TCP SPT=34264 DPT=9882 SEQ=1633358685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A128FF7970000000001030307) Nov 27 04:29:38 localhost python3.9[200386]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtnodedevd.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:39 localhost python3.9[200496]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtnodedevd-ro.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:40 localhost python3.9[200606]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtnodedevd-admin.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:29:40 localhost podman[200717]: 2025-11-27 09:29:40.60858764 +0000 UTC m=+0.087804077 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 04:29:40 localhost podman[200717]: 2025-11-27 09:29:40.650661621 +0000 UTC m=+0.129878158 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 04:29:40 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:29:40 localhost python3.9[200716]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtproxyd.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:41 localhost python3.9[200849]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtproxyd-ro.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39963 DF PROTO=TCP SPT=34264 DPT=9882 SEQ=1633358685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290039A0000000001030307) Nov 27 04:29:42 localhost python3.9[200959]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtproxyd-admin.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:42 localhost python3.9[201069]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtqemud.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:29:43.534 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:29:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:29:43.535 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:29:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:29:43.536 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:29:43 localhost python3.9[201179]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtqemud-ro.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:44 localhost python3.9[201289]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtqemud-admin.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:45 localhost python3.9[201399]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtsecretd.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25601 DF PROTO=TCP SPT=50938 DPT=9102 SEQ=1427925906 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129012290000000001030307) Nov 27 04:29:45 localhost python3.9[201509]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtsecretd-ro.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:46 localhost python3.9[201619]: ansible-ansible.builtin.file Invoked with group=root mode=0755 owner=root path=/etc/systemd/system/virtsecretd-admin.socket.d state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:29:47 localhost podman[201729]: 2025-11-27 09:29:47.241644334 +0000 UTC m=+0.098741677 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:29:47 localhost systemd[1]: tmp-crun.QbIG1Z.mount: Deactivated successfully. Nov 27 04:29:47 localhost podman[201729]: 2025-11-27 09:29:47.273785662 +0000 UTC m=+0.130883005 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_metadata_agent) Nov 27 04:29:47 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:29:47 localhost python3.9[201730]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtlogd.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22233 DF PROTO=TCP SPT=57054 DPT=9100 SEQ=2506281730 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12901A5A0000000001030307) Nov 27 04:29:47 localhost python3.9[201836]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtlogd.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235786.848366-2309-202861776122116/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:48 localhost python3.9[201946]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtlogd-admin.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:49 localhost python3.9[202034]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtlogd-admin.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235788.1273825-2309-213205099108464/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:49 localhost python3.9[202144]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtnodedevd.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:50 localhost python3.9[202232]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtnodedevd.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235789.3618975-2309-215334075782743/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:51 localhost python3.9[202342]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtnodedevd-ro.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:51 localhost python3.9[202430]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtnodedevd-ro.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235790.5689683-2309-122890461470324/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22234 DF PROTO=TCP SPT=57054 DPT=9100 SEQ=2506281730 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12902A1A0000000001030307) Nov 27 04:29:52 localhost python3.9[202540]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtnodedevd-admin.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:52 localhost python3.9[202628]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtnodedevd-admin.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235791.715227-2309-12557123500568/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:53 localhost python3.9[202738]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtproxyd.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31348 DF PROTO=TCP SPT=36320 DPT=9101 SEQ=2128586684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290321B0000000001030307) Nov 27 04:29:53 localhost python3.9[202826]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtproxyd.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235792.918639-2309-196912378012667/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:55 localhost python3.9[202936]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtproxyd-ro.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:55 localhost python3.9[203024]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtproxyd-ro.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235794.0713682-2309-192413073841832/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:56 localhost python3.9[203134]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtproxyd-admin.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=42397 DF PROTO=TCP SPT=39802 DPT=9101 SEQ=3226966282 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12903DDA0000000001030307) Nov 27 04:29:57 localhost python3.9[203222]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtproxyd-admin.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235796.0681496-2309-119201171302277/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:58 localhost python3.9[203332]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtqemud.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:29:59 localhost python3.9[203420]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtqemud.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235798.2702432-2309-209625431700350/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:29:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6043 DF PROTO=TCP SPT=34358 DPT=9105 SEQ=2862105774 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129049DA0000000001030307) Nov 27 04:29:59 localhost python3.9[203530]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtqemud-ro.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:00 localhost python3.9[203618]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtqemud-ro.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235799.399252-2309-194527664623167/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:01 localhost python3.9[203728]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtqemud-admin.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:01 localhost python3.9[203816]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtqemud-admin.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235800.5721486-2309-264819240637671/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:02 localhost python3.9[203926]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtsecretd.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:02 localhost python3.9[204014]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtsecretd.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235801.7686234-2309-38287363974693/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16994 DF PROTO=TCP SPT=42676 DPT=9105 SEQ=1984989856 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129055DB0000000001030307) Nov 27 04:30:03 localhost python3.9[204124]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtsecretd-ro.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:04 localhost python3.9[204212]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtsecretd-ro.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235802.9284463-2309-50194312363924/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:04 localhost python3.9[204322]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virtsecretd-admin.socket.d/override.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:05 localhost python3.9[204410]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virtsecretd-admin.socket.d/override.conf group=root mode=0644 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235804.1642087-2309-148959702117610/.source.conf follow=False _original_basename=libvirt-socket.unit.j2 checksum=0bad41f409b4ee7e780a2a59dc18f5c84ed99826 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:05 localhost python3.9[204518]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail#012ls -lRZ /run/libvirt | grep -E ':container_\S+_t'#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:06 localhost python3.9[204631]: ansible-ansible.posix.seboolean Invoked with name=os_enable_vtpm persistent=True state=True ignore_selinux_state=False Nov 27 04:30:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31351 DF PROTO=TCP SPT=36320 DPT=9101 SEQ=2128586684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129069DA0000000001030307) Nov 27 04:30:08 localhost python3.9[204741]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True name=virtlogd.service state=restarted daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:30:08 localhost systemd[1]: Reloading. Nov 27 04:30:08 localhost systemd-rc-local-generator[204768]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:08 localhost systemd-sysv-generator[204771]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:08 localhost systemd[1]: Starting libvirt logging daemon socket... Nov 27 04:30:08 localhost systemd[1]: Listening on libvirt logging daemon socket. Nov 27 04:30:08 localhost systemd[1]: Starting libvirt logging daemon admin socket... Nov 27 04:30:08 localhost systemd[1]: Listening on libvirt logging daemon admin socket. Nov 27 04:30:08 localhost systemd[1]: Starting libvirt logging daemon... Nov 27 04:30:08 localhost systemd[1]: Started libvirt logging daemon. Nov 27 04:30:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6325 DF PROTO=TCP SPT=59110 DPT=9882 SEQ=631309933 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12906CC70000000001030307) Nov 27 04:30:09 localhost python3.9[204893]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True name=virtnodedevd.service state=restarted daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:30:09 localhost systemd[1]: Reloading. Nov 27 04:30:09 localhost systemd-rc-local-generator[204915]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:09 localhost systemd-sysv-generator[204920]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:09 localhost systemd[1]: Starting libvirt nodedev daemon socket... Nov 27 04:30:09 localhost systemd[1]: Listening on libvirt nodedev daemon socket. Nov 27 04:30:09 localhost systemd[1]: Starting libvirt nodedev daemon admin socket... Nov 27 04:30:09 localhost systemd[1]: Starting libvirt nodedev daemon read-only socket... Nov 27 04:30:09 localhost systemd[1]: Listening on libvirt nodedev daemon admin socket. Nov 27 04:30:09 localhost systemd[1]: Listening on libvirt nodedev daemon read-only socket. Nov 27 04:30:09 localhost systemd[1]: Started libvirt nodedev daemon. Nov 27 04:30:10 localhost systemd[1]: Starting SETroubleshoot daemon for processing new SELinux denial logs... Nov 27 04:30:10 localhost systemd[1]: Started SETroubleshoot daemon for processing new SELinux denial logs. Nov 27 04:30:10 localhost setroubleshoot[205014]: Deleting alert 16fdbab6-fdc7-40ae-9244-67863b31fd5b, it is allowed in current policy Nov 27 04:30:10 localhost systemd[1]: Started dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@1.service. Nov 27 04:30:10 localhost python3.9[205069]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True name=virtproxyd.service state=restarted daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:30:10 localhost systemd[1]: Reloading. Nov 27 04:30:10 localhost systemd-sysv-generator[205107]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:10 localhost systemd-rc-local-generator[205104]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:30:10 localhost systemd[1]: Starting libvirt proxy daemon socket... Nov 27 04:30:10 localhost systemd[1]: Listening on libvirt proxy daemon socket. Nov 27 04:30:10 localhost systemd[1]: Starting libvirt proxy daemon admin socket... Nov 27 04:30:10 localhost systemd[1]: Starting libvirt proxy daemon read-only socket... Nov 27 04:30:10 localhost systemd[1]: Listening on libvirt proxy daemon admin socket. Nov 27 04:30:10 localhost systemd[1]: Listening on libvirt proxy daemon read-only socket. Nov 27 04:30:10 localhost systemd[1]: Started libvirt proxy daemon. Nov 27 04:30:11 localhost podman[205116]: 2025-11-27 09:30:11.02685625 +0000 UTC m=+0.080996408 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, managed_by=edpm_ansible) Nov 27 04:30:11 localhost podman[205116]: 2025-11-27 09:30:11.063304212 +0000 UTC m=+0.117444370 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125) Nov 27 04:30:11 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:30:11 localhost setroubleshoot[205014]: SELinux is preventing /usr/sbin/virtlogd from using the dac_read_search capability. For complete SELinux messages run: sealert -l 61d4893f-4019-452f-b874-a6a45599a8cc Nov 27 04:30:11 localhost setroubleshoot[205014]: SELinux is preventing /usr/sbin/virtlogd from using the dac_read_search capability.#012#012***** Plugin dac_override (91.4 confidence) suggests **********************#012#012If you want to help identify if domain needs this access or you have a file with the wrong permissions on your system#012Then turn on full auditing to get path information about the offending file and generate the error again.#012Do#012#012Turn on full auditing#012# auditctl -w /etc/shadow -p w#012Try to recreate AVC. Then execute#012# ausearch -m avc -ts recent#012If you see PATH record check ownership/permissions on file, and fix it,#012otherwise report as a bugzilla.#012#012***** Plugin catchall (9.59 confidence) suggests **************************#012#012If you believe that virtlogd should have the dac_read_search capability by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'virtlogd' --raw | audit2allow -M my-virtlogd#012# semodule -X 300 -i my-virtlogd.pp#012 Nov 27 04:30:11 localhost setroubleshoot[205014]: SELinux is preventing /usr/sbin/virtlogd from using the dac_read_search capability. For complete SELinux messages run: sealert -l 61d4893f-4019-452f-b874-a6a45599a8cc Nov 27 04:30:11 localhost setroubleshoot[205014]: SELinux is preventing /usr/sbin/virtlogd from using the dac_read_search capability.#012#012***** Plugin dac_override (91.4 confidence) suggests **********************#012#012If you want to help identify if domain needs this access or you have a file with the wrong permissions on your system#012Then turn on full auditing to get path information about the offending file and generate the error again.#012Do#012#012Turn on full auditing#012# auditctl -w /etc/shadow -p w#012Try to recreate AVC. Then execute#012# ausearch -m avc -ts recent#012If you see PATH record check ownership/permissions on file, and fix it,#012otherwise report as a bugzilla.#012#012***** Plugin catchall (9.59 confidence) suggests **************************#012#012If you believe that virtlogd should have the dac_read_search capability by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'virtlogd' --raw | audit2allow -M my-virtlogd#012# semodule -X 300 -i my-virtlogd.pp#012 Nov 27 04:30:11 localhost python3.9[205274]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True name=virtqemud.service state=restarted daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:30:11 localhost systemd[1]: Reloading. Nov 27 04:30:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=6327 DF PROTO=TCP SPT=59110 DPT=9882 SEQ=631309933 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129078DA0000000001030307) Nov 27 04:30:11 localhost systemd-rc-local-generator[205297]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:11 localhost systemd-sysv-generator[205303]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:11 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:12 localhost systemd[1]: Listening on libvirt locking daemon socket. Nov 27 04:30:12 localhost systemd[1]: Starting libvirt QEMU daemon socket... Nov 27 04:30:12 localhost systemd[1]: Listening on libvirt QEMU daemon socket. Nov 27 04:30:12 localhost systemd[1]: Starting libvirt QEMU daemon admin socket... Nov 27 04:30:12 localhost systemd[1]: Starting libvirt QEMU daemon read-only socket... Nov 27 04:30:12 localhost systemd[1]: Listening on libvirt QEMU daemon admin socket. Nov 27 04:30:12 localhost systemd[1]: Listening on libvirt QEMU daemon read-only socket. Nov 27 04:30:12 localhost systemd[1]: Started libvirt QEMU daemon. Nov 27 04:30:12 localhost python3.9[205455]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True name=virtsecretd.service state=restarted daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:30:12 localhost systemd[1]: Reloading. Nov 27 04:30:13 localhost systemd-sysv-generator[205529]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:13 localhost systemd-rc-local-generator[205522]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:13 localhost systemd[1]: Starting libvirt secret daemon socket... Nov 27 04:30:13 localhost systemd[1]: Listening on libvirt secret daemon socket. Nov 27 04:30:13 localhost systemd[1]: Starting libvirt secret daemon admin socket... Nov 27 04:30:13 localhost systemd[1]: Starting libvirt secret daemon read-only socket... Nov 27 04:30:13 localhost systemd[1]: Listening on libvirt secret daemon admin socket. Nov 27 04:30:13 localhost systemd[1]: Listening on libvirt secret daemon read-only socket. Nov 27 04:30:13 localhost systemd[1]: Started libvirt secret daemon. Nov 27 04:30:14 localhost python3.9[205705]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/openstack/config/ceph state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:14 localhost python3.9[205815]: ansible-ansible.builtin.find Invoked with paths=['/var/lib/openstack/config/ceph'] patterns=['*.conf'] read_whole_file=False file_type=file age_stamp=mtime recurse=False hidden=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:30:15 localhost python3.9[205925]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail;#012echo ceph#012awk -F '=' '/fsid/ {print $2}' /var/lib/openstack/config/ceph/ceph.conf | xargs#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=60953 DF PROTO=TCP SPT=44898 DPT=9102 SEQ=990281826 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129087590000000001030307) Nov 27 04:30:16 localhost python3.9[206037]: ansible-ansible.builtin.find Invoked with paths=['/var/lib/openstack/config/ceph'] patterns=['*.keyring'] read_whole_file=False file_type=file age_stamp=mtime recurse=False hidden=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:30:17 localhost python3.9[206164]: ansible-ansible.legacy.stat Invoked with path=/tmp/secret.xml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:17 localhost python3.9[206251]: ansible-ansible.legacy.copy Invoked with dest=/tmp/secret.xml mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235816.6171346-3173-170738949055151/.source.xml follow=False _original_basename=secret.xml.j2 checksum=631150486f483cf9bd5def30ceb046b83248bdba backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57416 DF PROTO=TCP SPT=45406 DPT=9100 SEQ=2535589314 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12908F9B0000000001030307) Nov 27 04:30:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:30:18 localhost podman[206323]: 2025-11-27 09:30:18.005833772 +0000 UTC m=+0.094045322 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:30:18 localhost podman[206323]: 2025-11-27 09:30:18.012204361 +0000 UTC m=+0.100415941 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_managed=true, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:30:18 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:30:18 localhost python3.9[206378]: ansible-ansible.legacy.command Invoked with _raw_params=virsh secret-undefine e83f3b0c-4090-52df-95d4-ad9be8516692#012virsh secret-define --file /tmp/secret.xml#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:19 localhost python3.9[206616]: ansible-ansible.builtin.file Invoked with path=/tmp/secret.xml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:21 localhost systemd[1]: dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@1.service: Deactivated successfully. Nov 27 04:30:21 localhost systemd[1]: setroubleshootd.service: Deactivated successfully. Nov 27 04:30:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57417 DF PROTO=TCP SPT=45406 DPT=9100 SEQ=2535589314 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12909F5B0000000001030307) Nov 27 04:30:21 localhost python3.9[207010]: ansible-ansible.legacy.copy Invoked with dest=/etc/ceph/ceph.conf group=root mode=0644 owner=root remote_src=True src=/var/lib/openstack/config/ceph/ceph.conf backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:23 localhost python3.9[207120]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/libvirt.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55284 DF PROTO=TCP SPT=60418 DPT=9101 SEQ=2075428807 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290A71E0000000001030307) Nov 27 04:30:24 localhost python3.9[207208]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/edpm-config/firewall/libvirt.yaml mode=0640 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235822.2449207-3338-195719488883181/.source.yaml follow=False _original_basename=firewall.yaml.j2 checksum=dc5ee7162311c27a6084cbee4052b901d56cb1ba backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:25 localhost python3.9[207318]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/var/lib/edpm-config/firewall state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:25 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 04:30:25 localhost rhsm-service[6591]: WARNING [subscription_manager.cert_sorter:194] Installed product 479 not present in response from server. Nov 27 04:30:25 localhost python3.9[207430]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:26 localhost python3.9[207487]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml _original_basename=base-rules.yaml.j2 recurse=False state=file path=/var/lib/edpm-config/firewall/edpm-nftables-base.yaml force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:26 localhost python3.9[207597]: ansible-ansible.legacy.stat Invoked with path=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46907 DF PROTO=TCP SPT=49890 DPT=9105 SEQ=18843434 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290B35A0000000001030307) Nov 27 04:30:27 localhost python3.9[207654]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml _original_basename=.ped_au5w recurse=False state=file path=/var/lib/edpm-config/firewall/edpm-nftables-user-rules.yaml force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:27 localhost python3.9[207764]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/iptables.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:28 localhost python3.9[207821]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/iptables.nft _original_basename=iptables.nft recurse=False state=file path=/etc/nftables/iptables.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:29 localhost python3.9[207931]: ansible-ansible.legacy.command Invoked with _raw_params=nft -j list ruleset _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55286 DF PROTO=TCP SPT=60418 DPT=9101 SEQ=2075428807 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290BEDA0000000001030307) Nov 27 04:30:29 localhost python3[208042]: ansible-edpm_nftables_from_files Invoked with src=/var/lib/edpm-config/firewall Nov 27 04:30:30 localhost python3.9[208152]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:31 localhost python3.9[208209]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/edpm-jumps.nft _original_basename=jump-chain.j2 recurse=False state=file path=/etc/nftables/edpm-jumps.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:31 localhost python3.9[208319]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-update-jumps.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:32 localhost python3.9[208376]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/edpm-update-jumps.nft _original_basename=jump-chain.j2 recurse=False state=file path=/etc/nftables/edpm-update-jumps.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46909 DF PROTO=TCP SPT=49890 DPT=9105 SEQ=18843434 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290CB1B0000000001030307) Nov 27 04:30:33 localhost python3.9[208486]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-flushes.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:34 localhost python3.9[208543]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/edpm-flushes.nft _original_basename=flush-chain.j2 recurse=False state=file path=/etc/nftables/edpm-flushes.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:34 localhost python3.9[208653]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-chains.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:35 localhost python3.9[208710]: ansible-ansible.legacy.file Invoked with group=root mode=0600 owner=root dest=/etc/nftables/edpm-chains.nft _original_basename=chains.j2 recurse=False state=file path=/etc/nftables/edpm-chains.nft force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:36 localhost python3.9[208820]: ansible-ansible.legacy.stat Invoked with path=/etc/nftables/edpm-rules.nft follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:37 localhost python3.9[208910]: ansible-ansible.legacy.copy Invoked with dest=/etc/nftables/edpm-rules.nft group=root mode=0600 owner=root src=/home/zuul/.ansible/tmp/ansible-tmp-1764235836.1398067-3713-219620874836171/.source.nft follow=False _original_basename=ruleset.j2 checksum=e2e2635f27347d386f310e86d2b40c40289835bb backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:38 localhost python3.9[209020]: ansible-ansible.builtin.file Invoked with group=root mode=0600 owner=root path=/etc/nftables/edpm-rules.nft.changed state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55287 DF PROTO=TCP SPT=60418 DPT=9101 SEQ=2075428807 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290DFDA0000000001030307) Nov 27 04:30:38 localhost python3.9[209130]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-chains.nft /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft /etc/nftables/edpm-jumps.nft | nft -c -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14278 DF PROTO=TCP SPT=50736 DPT=9882 SEQ=1196179700 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290E1F70000000001030307) Nov 27 04:30:39 localhost python3.9[209243]: ansible-ansible.builtin.blockinfile Invoked with backup=False block=include "/etc/nftables/iptables.nft"#012include "/etc/nftables/edpm-chains.nft"#012include "/etc/nftables/edpm-rules.nft"#012include "/etc/nftables/edpm-jumps.nft"#012 path=/etc/sysconfig/nftables.conf validate=nft -c -f %s state=present marker=# {mark} ANSIBLE MANAGED BLOCK create=False marker_begin=BEGIN marker_end=END append_newline=False prepend_newline=False encoding=utf-8 unsafe_writes=False insertafter=None insertbefore=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:40 localhost python3.9[209353]: ansible-ansible.legacy.command Invoked with _raw_params=nft -f /etc/nftables/edpm-chains.nft _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:40 localhost python3.9[209464]: ansible-ansible.builtin.stat Invoked with path=/etc/nftables/edpm-rules.nft.changed follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:30:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:30:41 localhost podman[209577]: 2025-11-27 09:30:41.559122771 +0000 UTC m=+0.074482323 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:30:41 localhost podman[209577]: 2025-11-27 09:30:41.593223991 +0000 UTC m=+0.108583543 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_controller) Nov 27 04:30:41 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:30:41 localhost python3.9[209576]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; cat /etc/nftables/edpm-flushes.nft /etc/nftables/edpm-rules.nft /etc/nftables/edpm-update-jumps.nft | nft -f - _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:30:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=14280 DF PROTO=TCP SPT=50736 DPT=9882 SEQ=1196179700 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290EE1A0000000001030307) Nov 27 04:30:42 localhost python3.9[209714]: ansible-ansible.builtin.file Invoked with path=/etc/nftables/edpm-rules.nft.changed state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:43 localhost python3.9[209824]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm_libvirt.target follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:30:43.535 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:30:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:30:43.535 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:30:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:30:43.537 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:30:43 localhost python3.9[209912]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/edpm_libvirt.target mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235842.6016095-3929-178600772738291/.source.target follow=False _original_basename=edpm_libvirt.target checksum=13035a1aa0f414c677b14be9a5a363b6623d393c backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:44 localhost python3.9[210022]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm_libvirt_guests.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:44 localhost python3.9[210110]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/edpm_libvirt_guests.service mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235843.839417-3974-200643789459415/.source.service follow=False _original_basename=edpm_libvirt_guests.service checksum=db83430a42fc2ccfd6ed8b56ebf04f3dff9cd0cf backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:45 localhost python3.9[210220]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/virt-guest-shutdown.target follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:30:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28539 DF PROTO=TCP SPT=50042 DPT=9102 SEQ=4130918596 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1290FC890000000001030307) Nov 27 04:30:46 localhost python3.9[210308]: ansible-ansible.legacy.copy Invoked with dest=/etc/systemd/system/virt-guest-shutdown.target mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235845.0712447-4019-221969333024685/.source.target follow=False _original_basename=virt-guest-shutdown.target checksum=49ca149619c596cbba877418629d2cf8f7b0f5cf backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:30:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27870 DF PROTO=TCP SPT=50626 DPT=9100 SEQ=3345989524 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291049A0000000001030307) Nov 27 04:30:47 localhost python3.9[210418]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm_libvirt.target state=restarted daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:30:47 localhost systemd[1]: Reloading. Nov 27 04:30:47 localhost systemd-rc-local-generator[210445]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:47 localhost systemd-sysv-generator[210448]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:47 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:48 localhost systemd[1]: Reached target edpm_libvirt.target. Nov 27 04:30:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:30:49 localhost systemd[1]: tmp-crun.7tOiCJ.mount: Deactivated successfully. Nov 27 04:30:49 localhost podman[210475]: 2025-11-27 09:30:49.018067215 +0000 UTC m=+0.107668849 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:30:49 localhost podman[210475]: 2025-11-27 09:30:49.0509127 +0000 UTC m=+0.140514334 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:30:49 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:30:49 localhost python3.9[210583]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm_libvirt_guests daemon_reexec=False scope=system no_block=False state=None force=None masked=None Nov 27 04:30:49 localhost systemd[1]: Reloading. Nov 27 04:30:49 localhost systemd-rc-local-generator[210606]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:49 localhost systemd-sysv-generator[210612]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:49 localhost systemd[1]: Reloading. Nov 27 04:30:50 localhost systemd-rc-local-generator[210648]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:30:50 localhost systemd-sysv-generator[210651]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:30:50 localhost systemd[1]: session-53.scope: Deactivated successfully. Nov 27 04:30:50 localhost systemd[1]: session-53.scope: Consumed 3min 41.582s CPU time. Nov 27 04:30:50 localhost systemd-logind[761]: Session 53 logged out. Waiting for processes to exit. Nov 27 04:30:50 localhost systemd-logind[761]: Removed session 53. Nov 27 04:30:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27871 DF PROTO=TCP SPT=50626 DPT=9100 SEQ=3345989524 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291145A0000000001030307) Nov 27 04:30:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18369 DF PROTO=TCP SPT=46742 DPT=9101 SEQ=866792162 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12911C5A0000000001030307) Nov 27 04:30:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31353 DF PROTO=TCP SPT=36320 DPT=9101 SEQ=2128586684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129127DB0000000001030307) Nov 27 04:30:56 localhost sshd[210674]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:30:56 localhost systemd-logind[761]: New session 54 of user zuul. Nov 27 04:30:56 localhost systemd[1]: Started Session 54 of User zuul. Nov 27 04:30:57 localhost python3.9[210785]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:30:59 localhost python3.9[210897]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:30:59 localhost network[210914]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:30:59 localhost network[210915]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:30:59 localhost network[210916]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:30:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=27872 DF PROTO=TCP SPT=50626 DPT=9100 SEQ=3345989524 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129133DA0000000001030307) Nov 27 04:31:00 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:31:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9883 DF PROTO=TCP SPT=40158 DPT=9105 SEQ=232022366 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291405B0000000001030307) Nov 27 04:31:04 localhost python3.9[211148]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 27 04:31:05 localhost python3.9[211211]: ansible-ansible.legacy.dnf Invoked with name=['iscsi-initiator-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:31:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=18372 DF PROTO=TCP SPT=46742 DPT=9101 SEQ=866792162 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129153DA0000000001030307) Nov 27 04:31:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24143 DF PROTO=TCP SPT=53464 DPT=9882 SEQ=3195835258 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129157270000000001030307) Nov 27 04:31:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24145 DF PROTO=TCP SPT=53464 DPT=9882 SEQ=3195835258 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291631A0000000001030307) Nov 27 04:31:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:31:11 localhost podman[211231]: 2025-11-27 09:31:11.991782452 +0000 UTC m=+0.081726375 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:31:12 localhost podman[211231]: 2025-11-27 09:31:12.06101058 +0000 UTC m=+0.150954453 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:31:12 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:31:13 localhost python3.9[211351]: ansible-ansible.builtin.stat Invoked with path=/var/lib/config-data/puppet-generated/iscsid/etc/iscsi follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:31:14 localhost python3.9[211463]: ansible-ansible.legacy.copy Invoked with dest=/etc/iscsi mode=preserve remote_src=True src=/var/lib/config-data/puppet-generated/iscsid/etc/iscsi/ backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4166 DF PROTO=TCP SPT=45494 DPT=9102 SEQ=1037592795 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129171B90000000001030307) Nov 27 04:31:16 localhost python3.9[211573]: ansible-ansible.legacy.command Invoked with _raw_params=mv "/var/lib/config-data/puppet-generated/iscsid/etc/iscsi" "/var/lib/config-data/puppet-generated/iscsid/etc/iscsi.adopted"#012 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:31:16 localhost python3.9[211720]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/restorecon -nvr /etc/iscsi /var/lib/iscsi _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:31:17 localhost python3.9[211888]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/restorecon -rF /etc/iscsi /var/lib/iscsi _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:31:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38248 DF PROTO=TCP SPT=42210 DPT=9100 SEQ=1301178173 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129179DA0000000001030307) Nov 27 04:31:18 localhost python3.9[212032]: ansible-ansible.builtin.stat Invoked with path=/etc/iscsi/.initiator_reset follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:31:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:31:19 localhost systemd[1]: tmp-crun.rEN43X.mount: Deactivated successfully. Nov 27 04:31:19 localhost podman[212163]: 2025-11-27 09:31:19.482495701 +0000 UTC m=+0.092945774 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:31:19 localhost podman[212163]: 2025-11-27 09:31:19.516987592 +0000 UTC m=+0.127437665 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:31:19 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:31:19 localhost python3.9[212162]: ansible-ansible.builtin.lineinfile Invoked with insertafter=^#node.session.auth.chap.algs line=node.session.auth.chap_algs = SHA3-256,SHA256,SHA1,MD5 path=/etc/iscsi/iscsid.conf regexp=^node.session.auth.chap_algs state=present encoding=utf-8 backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:20 localhost python3.9[212290]: ansible-ansible.builtin.systemd_service Invoked with enabled=True name=iscsid.socket state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:31:20 localhost systemd[1]: Listening on Open-iSCSI iscsid Socket. Nov 27 04:31:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38249 DF PROTO=TCP SPT=42210 DPT=9100 SEQ=1301178173 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291899A0000000001030307) Nov 27 04:31:21 localhost python3.9[212404]: ansible-ansible.builtin.systemd_service Invoked with enabled=True name=iscsid state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:31:21 localhost systemd[1]: Reloading. Nov 27 04:31:21 localhost systemd-rc-local-generator[212427]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:31:21 localhost systemd-sysv-generator[212437]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:31:21 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:21 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:21 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:21 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:21 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:31:22 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:22 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:22 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:22 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:22 localhost systemd[1]: One time configuration for iscsi.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/iscsi/initiatorname.iscsi). Nov 27 04:31:22 localhost systemd[1]: Starting Open-iSCSI... Nov 27 04:31:22 localhost iscsid[212446]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Nov 27 04:31:22 localhost iscsid[212446]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Nov 27 04:31:22 localhost iscsid[212446]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Nov 27 04:31:22 localhost iscsid[212446]: If using hardware iscsi like qla4xxx this message can be ignored. Nov 27 04:31:22 localhost iscsid[212446]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Nov 27 04:31:22 localhost iscsid[212446]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Nov 27 04:31:22 localhost iscsid[212446]: iscsid: can't open iscsid.ipc_auth_uid configuration file /etc/iscsi/iscsid.conf Nov 27 04:31:22 localhost systemd[1]: Started Open-iSCSI. Nov 27 04:31:22 localhost systemd[1]: Starting Logout off all iSCSI sessions on shutdown... Nov 27 04:31:22 localhost systemd[1]: Finished Logout off all iSCSI sessions on shutdown. Nov 27 04:31:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31052 DF PROTO=TCP SPT=45436 DPT=9101 SEQ=1787421661 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291919A0000000001030307) Nov 27 04:31:24 localhost systemd[1]: Starting SETroubleshoot daemon for processing new SELinux denial logs... Nov 27 04:31:24 localhost python3.9[212557]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:31:24 localhost network[212575]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:31:24 localhost network[212576]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:31:24 localhost network[212577]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:31:24 localhost systemd[1]: Started SETroubleshoot daemon for processing new SELinux denial logs. Nov 27 04:31:24 localhost systemd[1]: Started dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@2.service. Nov 27 04:31:25 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi. For complete SELinux messages run: sealert -l 138fa0e0-9db4-4dde-b634-71550f6cd433 Nov 27 04:31:25 localhost setroubleshoot[212558]: SELinux is preventing /usr/sbin/iscsid from search access on the directory iscsi.#012#012***** Plugin catchall (100. confidence) suggests **************************#012#012If you believe that iscsid should be allowed search access on the iscsi directory by default.#012Then you should report this as a bug.#012You can generate a local policy module to allow this access.#012Do#012allow this access for now by executing:#012# ausearch -c 'iscsid' --raw | audit2allow -M my-iscsid#012# semodule -X 300 -i my-iscsid.pp#012 Nov 27 04:31:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44538 DF PROTO=TCP SPT=53772 DPT=9105 SEQ=866799374 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12919D9B0000000001030307) Nov 27 04:31:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31054 DF PROTO=TCP SPT=45436 DPT=9101 SEQ=1787421661 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291A95B0000000001030307) Nov 27 04:31:30 localhost python3.9[212824]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/modules-load.d selevel=s0 setype=etc_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 04:31:31 localhost sshd[212934]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:31:31 localhost python3.9[212935]: ansible-community.general.modprobe Invoked with name=dm-multipath state=present params= persistent=disabled Nov 27 04:31:32 localhost python3.9[213050]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/dm-multipath.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:32 localhost python3.9[213138]: ansible-ansible.legacy.copy Invoked with dest=/etc/modules-load.d/dm-multipath.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235891.7777648-458-235434936860852/.source.conf follow=False _original_basename=module-load.conf.j2 checksum=065061c60917e4f67cecc70d12ce55e42f9d0b3f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44540 DF PROTO=TCP SPT=53772 DPT=9105 SEQ=866799374 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291B55A0000000001030307) Nov 27 04:31:33 localhost python3.9[213248]: ansible-ansible.builtin.lineinfile Invoked with create=True dest=/etc/modules line=dm-multipath mode=0644 state=present path=/etc/modules encoding=utf-8 backrefs=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:34 localhost python3.9[213358]: ansible-ansible.builtin.systemd Invoked with name=systemd-modules-load.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:31:34 localhost systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 27 04:31:34 localhost systemd[1]: Stopped Load Kernel Modules. Nov 27 04:31:34 localhost systemd[1]: Stopping Load Kernel Modules... Nov 27 04:31:34 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 04:31:34 localhost systemd-modules-load[213362]: Module 'msr' is built in Nov 27 04:31:34 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 04:31:35 localhost python3.9[213472]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/multipath setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:31:35 localhost systemd[1]: dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@2.service: Deactivated successfully. Nov 27 04:31:35 localhost systemd[1]: dbus-:1.1-org.fedoraproject.SetroubleshootPrivileged@2.service: Consumed 1.006s CPU time. Nov 27 04:31:35 localhost systemd[1]: setroubleshootd.service: Deactivated successfully. Nov 27 04:31:36 localhost python3.9[213582]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:31:37 localhost python3.9[213692]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:31:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31055 DF PROTO=TCP SPT=45436 DPT=9101 SEQ=1787421661 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291C9DA0000000001030307) Nov 27 04:31:38 localhost python3.9[213802]: ansible-ansible.legacy.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:38 localhost python3.9[213890]: ansible-ansible.legacy.copy Invoked with dest=/etc/multipath.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235897.65734-631-37634507430614/.source.conf _original_basename=multipath.conf follow=False checksum=bf02ab264d3d648048a81f3bacec8bc58db93162 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40476 DF PROTO=TCP SPT=43746 DPT=9882 SEQ=3508207314 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291CC570000000001030307) Nov 27 04:31:40 localhost python3.9[214000]: ansible-ansible.legacy.command Invoked with _raw_params=grep -q '^blacklist\s*{' /etc/multipath.conf _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:31:40 localhost python3.9[214111]: ansible-ansible.builtin.lineinfile Invoked with line=blacklist { path=/etc/multipath.conf state=present encoding=utf-8 backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40478 DF PROTO=TCP SPT=43746 DPT=9882 SEQ=3508207314 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291D85A0000000001030307) Nov 27 04:31:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:31:42 localhost podman[214222]: 2025-11-27 09:31:42.426425053 +0000 UTC m=+0.083360124 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_controller) Nov 27 04:31:42 localhost podman[214222]: 2025-11-27 09:31:42.498925059 +0000 UTC m=+0.155860140 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:31:42 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:31:42 localhost python3.9[214221]: ansible-ansible.builtin.replace Invoked with path=/etc/multipath.conf regexp=^(blacklist {) replace=\1\n} backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:43 localhost python3.9[214355]: ansible-ansible.builtin.replace Invoked with path=/etc/multipath.conf regexp=^blacklist\s*{\n[\s]+devnode \"\.\*\" replace=blacklist { backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:31:43.537 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:31:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:31:43.538 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:31:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:31:43.539 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:31:43 localhost python3.9[214465]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= find_multipaths yes path=/etc/multipath.conf regexp=^\s+find_multipaths state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:44 localhost python3.9[214575]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= recheck_wwid yes path=/etc/multipath.conf regexp=^\s+recheck_wwid state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:45 localhost python3.9[214685]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= skip_kpartx yes path=/etc/multipath.conf regexp=^\s+skip_kpartx state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=57799 DF PROTO=TCP SPT=34840 DPT=9102 SEQ=1664626935 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291E6E90000000001030307) Nov 27 04:31:45 localhost python3.9[214795]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= user_friendly_names no path=/etc/multipath.conf regexp=^\s+user_friendly_names state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:46 localhost python3.9[214905]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:31:47 localhost python3.9[215017]: ansible-ansible.builtin.file Invoked with mode=0644 path=/etc/multipath/.multipath_restart_required state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=56429 DF PROTO=TCP SPT=38536 DPT=9100 SEQ=3345096170 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291EF1A0000000001030307) Nov 27 04:31:48 localhost python3.9[215127]: ansible-ansible.builtin.file Invoked with path=/var/local/libexec recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:31:49 localhost python3.9[215237]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-container-shutdown follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:49 localhost python3.9[215294]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-container-shutdown _original_basename=edpm-container-shutdown recurse=False state=file path=/var/local/libexec/edpm-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:31:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:31:49 localhost podman[215371]: 2025-11-27 09:31:49.990638347 +0000 UTC m=+0.085142013 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2) Nov 27 04:31:49 localhost podman[215371]: 2025-11-27 09:31:49.994164187 +0000 UTC m=+0.088667903 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:31:50 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:31:50 localhost python3.9[215422]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-start-podman-container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:50 localhost python3.9[215479]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-start-podman-container _original_basename=edpm-start-podman-container recurse=False state=file path=/var/local/libexec/edpm-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:31:51 localhost python3.9[215589]: ansible-ansible.builtin.file Invoked with mode=420 path=/etc/systemd/system-preset state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=56430 DF PROTO=TCP SPT=38536 DPT=9100 SEQ=3345096170 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1291FEDC0000000001030307) Nov 27 04:31:51 localhost python3.9[215699]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm-container-shutdown.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:53 localhost python3.9[215756]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/edpm-container-shutdown.service _original_basename=edpm-container-shutdown-service recurse=False state=file path=/etc/systemd/system/edpm-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37769 DF PROTO=TCP SPT=49092 DPT=9101 SEQ=1270208401 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129206DA0000000001030307) Nov 27 04:31:53 localhost python3.9[215866]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:54 localhost python3.9[215923]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-edpm-container-shutdown.preset _original_basename=91-edpm-container-shutdown-preset recurse=False state=file path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:55 localhost python3.9[216033]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm-container-shutdown state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:31:55 localhost systemd[1]: Reloading. Nov 27 04:31:55 localhost systemd-rc-local-generator[216060]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:31:55 localhost systemd-sysv-generator[216064]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:55 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20454 DF PROTO=TCP SPT=43388 DPT=9105 SEQ=1353290608 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129212DA0000000001030307) Nov 27 04:31:57 localhost python3.9[216181]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/netns-placeholder.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:57 localhost python3.9[216238]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/etc/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:58 localhost python3.9[216348]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:31:58 localhost python3.9[216405]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/etc/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:31:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9886 DF PROTO=TCP SPT=40158 DPT=9105 SEQ=232022366 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12921DDA0000000001030307) Nov 27 04:31:59 localhost python3.9[216515]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=netns-placeholder state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:31:59 localhost systemd[1]: Reloading. Nov 27 04:31:59 localhost systemd-sysv-generator[216546]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:31:59 localhost systemd-rc-local-generator[216542]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:31:59 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:00 localhost systemd[1]: Starting Create netns directory... Nov 27 04:32:00 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 04:32:00 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 04:32:00 localhost systemd[1]: Finished Create netns directory. Nov 27 04:32:01 localhost python3.9[216667]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/healthchecks setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:32:02 localhost python3.9[216777]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/multipathd/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:32:02 localhost python3.9[216865]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/multipathd/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764235921.6490579-1254-236995984841735/.source _original_basename=healthcheck follow=False checksum=af9d0c1c8f3cb0e30ce9609be9d5b01924d0d23f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:32:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20456 DF PROTO=TCP SPT=43388 DPT=9105 SEQ=1353290608 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12922A9B0000000001030307) Nov 27 04:32:03 localhost python3.9[216975]: ansible-ansible.builtin.file Invoked with path=/var/lib/kolla/config_files recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:32:04 localhost python3.9[217085]: ansible-ansible.legacy.stat Invoked with path=/var/lib/kolla/config_files/multipathd.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:32:04 localhost python3.9[217173]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/kolla/config_files/multipathd.json mode=0600 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235923.891108-1327-108260437639418/.source.json _original_basename=.4l10yvpd follow=False checksum=3f7959ee8ac9757398adcc451c3b416c957d7c14 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:06 localhost python3.9[217283]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/edpm-config/container-startup-config/multipathd state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37772 DF PROTO=TCP SPT=49092 DPT=9101 SEQ=1270208401 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12923FDA0000000001030307) Nov 27 04:32:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35644 DF PROTO=TCP SPT=50828 DPT=9882 SEQ=1803204025 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129241870000000001030307) Nov 27 04:32:09 localhost python3.9[217591]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/edpm-config/container-startup-config/multipathd config_pattern=*.json debug=False Nov 27 04:32:09 localhost systemd[1]: virtnodedevd.service: Deactivated successfully. Nov 27 04:32:10 localhost python3.9[217702]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:32:11 localhost systemd[1]: virtproxyd.service: Deactivated successfully. Nov 27 04:32:11 localhost python3.9[217813]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Nov 27 04:32:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35646 DF PROTO=TCP SPT=50828 DPT=9882 SEQ=1803204025 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12924D9B0000000001030307) Nov 27 04:32:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:32:12 localhost systemd[1]: tmp-crun.Szz7uu.mount: Deactivated successfully. Nov 27 04:32:13 localhost podman[217858]: 2025-11-27 09:32:13.001563696 +0000 UTC m=+0.089935420 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:32:13 localhost podman[217858]: 2025-11-27 09:32:13.036396258 +0000 UTC m=+0.124767972 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 04:32:13 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:32:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=46648 DF PROTO=TCP SPT=44400 DPT=9102 SEQ=3020186264 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12925C190000000001030307) Nov 27 04:32:16 localhost python3[217976]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/edpm-config/container-startup-config/multipathd config_id=multipathd config_overrides={} config_patterns=*.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:32:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20979 DF PROTO=TCP SPT=42448 DPT=9100 SEQ=3952142204 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292645A0000000001030307) Nov 27 04:32:18 localhost podman[217989]: 2025-11-27 09:32:16.388622164 +0000 UTC m=+0.045695500 image pull quay.io/podified-antelope-centos9/openstack-multipathd:current-podified Nov 27 04:32:18 localhost podman[218037]: Nov 27 04:32:18 localhost podman[218037]: 2025-11-27 09:32:18.627657433 +0000 UTC m=+0.078177288 container create d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd) Nov 27 04:32:18 localhost podman[218037]: 2025-11-27 09:32:18.588344813 +0000 UTC m=+0.038864718 image pull quay.io/podified-antelope-centos9/openstack-multipathd:current-podified Nov 27 04:32:18 localhost python3[217976]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name multipathd --conmon-pidfile /run/multipathd.pid --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --healthcheck-command /openstack/healthcheck --label config_id=multipathd --label container_name=multipathd --label managed_by=edpm_ansible --label config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']} --log-driver journald --log-level info --network host --privileged=True --volume /etc/hosts:/etc/hosts:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro --volume /dev/log:/dev/log --volume /var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro --volume /dev:/dev --volume /run/udev:/run/udev --volume /sys:/sys --volume /lib/modules:/lib/modules:ro --volume /etc/iscsi:/etc/iscsi:ro --volume /var/lib/iscsi:/var/lib/iscsi --volume /etc/multipath:/etc/multipath:z --volume /etc/multipath.conf:/etc/multipath.conf:ro --volume /var/lib/openstack/healthchecks/multipathd:/openstack:ro,z quay.io/podified-antelope-centos9/openstack-multipathd:current-podified Nov 27 04:32:19 localhost python3.9[218248]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:32:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:32:20 localhost podman[218286]: 2025-11-27 09:32:20.280655283 +0000 UTC m=+0.082357815 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:32:20 localhost podman[218286]: 2025-11-27 09:32:20.29186531 +0000 UTC m=+0.093567852 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent) Nov 27 04:32:20 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:32:20 localhost systemd[1]: virtsecretd.service: Deactivated successfully. Nov 27 04:32:21 localhost python3.9[218397]: ansible-file Invoked with path=/etc/systemd/system/edpm_multipathd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:21 localhost python3.9[218452]: ansible-stat Invoked with path=/etc/systemd/system/edpm_multipathd_healthcheck.timer follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:32:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20980 DF PROTO=TCP SPT=42448 DPT=9100 SEQ=3952142204 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292741A0000000001030307) Nov 27 04:32:22 localhost python3.9[218561]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764235941.6473713-1591-244506180447350/source dest=/etc/systemd/system/edpm_multipathd.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:22 localhost python3.9[218616]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:32:22 localhost systemd[1]: Reloading. Nov 27 04:32:22 localhost systemd-sysv-generator[218642]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:22 localhost systemd-rc-local-generator[218639]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:22 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:22 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:22 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:22 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:22 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:23 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:23 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:23 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:23 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25651 DF PROTO=TCP SPT=34830 DPT=9101 SEQ=3648322424 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12927BDA0000000001030307) Nov 27 04:32:24 localhost python3.9[218706]: ansible-systemd Invoked with state=restarted name=edpm_multipathd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:24 localhost systemd[1]: Reloading. Nov 27 04:32:24 localhost systemd-sysv-generator[218738]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:24 localhost systemd-rc-local-generator[218735]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:24 localhost systemd[1]: Starting multipathd container... Nov 27 04:32:24 localhost systemd[1]: Started libcrun container. Nov 27 04:32:24 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1e88cafbea01bedcfd0bc7269e8c324f9e046bf7e85136e02fdd1f9176fa84ad/merged/etc/multipath supports timestamps until 2038 (0x7fffffff) Nov 27 04:32:24 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1e88cafbea01bedcfd0bc7269e8c324f9e046bf7e85136e02fdd1f9176fa84ad/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 04:32:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:24 localhost podman[218747]: 2025-11-27 09:32:24.598670046 +0000 UTC m=+0.154435810 container init d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 04:32:24 localhost multipathd[218761]: + sudo -E kolla_set_configs Nov 27 04:32:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:24 localhost podman[218747]: 2025-11-27 09:32:24.64273891 +0000 UTC m=+0.198504654 container start d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_managed=true, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:32:24 localhost podman[218747]: multipathd Nov 27 04:32:24 localhost systemd[1]: Started multipathd container. Nov 27 04:32:24 localhost multipathd[218761]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:32:24 localhost multipathd[218761]: INFO:__main__:Validating config file Nov 27 04:32:24 localhost multipathd[218761]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:32:24 localhost multipathd[218761]: INFO:__main__:Writing out command to execute Nov 27 04:32:24 localhost multipathd[218761]: ++ cat /run_command Nov 27 04:32:24 localhost multipathd[218761]: + CMD='/usr/sbin/multipathd -d' Nov 27 04:32:24 localhost multipathd[218761]: + ARGS= Nov 27 04:32:24 localhost multipathd[218761]: + sudo kolla_copy_cacerts Nov 27 04:32:24 localhost multipathd[218761]: Running command: '/usr/sbin/multipathd -d' Nov 27 04:32:24 localhost multipathd[218761]: + [[ ! -n '' ]] Nov 27 04:32:24 localhost multipathd[218761]: + . kolla_extend_start Nov 27 04:32:24 localhost multipathd[218761]: + echo 'Running command: '\''/usr/sbin/multipathd -d'\''' Nov 27 04:32:24 localhost multipathd[218761]: + umask 0022 Nov 27 04:32:24 localhost multipathd[218761]: + exec /usr/sbin/multipathd -d Nov 27 04:32:24 localhost podman[218770]: 2025-11-27 09:32:24.724461506 +0000 UTC m=+0.075833111 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=starting, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:32:24 localhost multipathd[218761]: 10174.917487 | --------start up-------- Nov 27 04:32:24 localhost multipathd[218761]: 10174.917527 | read /etc/multipath.conf Nov 27 04:32:24 localhost multipathd[218761]: 10174.921214 | path checkers start up Nov 27 04:32:24 localhost podman[218770]: 2025-11-27 09:32:24.761927053 +0000 UTC m=+0.113298648 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, config_id=multipathd, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 04:32:24 localhost podman[218770]: unhealthy Nov 27 04:32:24 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:32:24 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Failed with result 'exit-code'. Nov 27 04:32:25 localhost python3.9[218907]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath/.multipath_restart_required follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:32:25 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.1 (250 of 333 items), suggesting rotation. Nov 27 04:32:25 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:32:25 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:32:25 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:32:26 localhost python3.9[219020]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --filter volume=/etc/multipath.conf --format {{.Names}} _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:32:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31057 DF PROTO=TCP SPT=45436 DPT=9101 SEQ=1787421661 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129287DA0000000001030307) Nov 27 04:32:27 localhost python3.9[219143]: ansible-ansible.builtin.systemd Invoked with name=edpm_multipathd state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:32:27 localhost systemd[1]: Stopping multipathd container... Nov 27 04:32:27 localhost systemd[1]: tmp-crun.1TpQDC.mount: Deactivated successfully. Nov 27 04:32:27 localhost multipathd[218761]: 10177.363396 | exit (signal) Nov 27 04:32:27 localhost multipathd[218761]: 10177.363463 | --------shut down------- Nov 27 04:32:27 localhost systemd[1]: libpod-d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.scope: Deactivated successfully. Nov 27 04:32:27 localhost podman[219147]: 2025-11-27 09:32:27.208023676 +0000 UTC m=+0.099048966 container died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=multipathd) Nov 27 04:32:27 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.timer: Deactivated successfully. Nov 27 04:32:27 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:27 localhost systemd[1]: tmp-crun.mABzR1.mount: Deactivated successfully. Nov 27 04:32:27 localhost podman[219147]: 2025-11-27 09:32:27.362379222 +0000 UTC m=+0.253404512 container cleanup d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:32:27 localhost podman[219147]: multipathd Nov 27 04:32:27 localhost podman[219176]: 2025-11-27 09:32:27.459616246 +0000 UTC m=+0.067776784 container cleanup d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:32:27 localhost podman[219176]: multipathd Nov 27 04:32:27 localhost systemd[1]: edpm_multipathd.service: Deactivated successfully. Nov 27 04:32:27 localhost systemd[1]: Stopped multipathd container. Nov 27 04:32:27 localhost systemd[1]: Starting multipathd container... Nov 27 04:32:27 localhost systemd[1]: Started libcrun container. Nov 27 04:32:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1e88cafbea01bedcfd0bc7269e8c324f9e046bf7e85136e02fdd1f9176fa84ad/merged/etc/multipath supports timestamps until 2038 (0x7fffffff) Nov 27 04:32:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1e88cafbea01bedcfd0bc7269e8c324f9e046bf7e85136e02fdd1f9176fa84ad/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 04:32:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:27 localhost podman[219189]: 2025-11-27 09:32:27.627492144 +0000 UTC m=+0.138165000 container init d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, container_name=multipathd) Nov 27 04:32:27 localhost multipathd[219203]: + sudo -E kolla_set_configs Nov 27 04:32:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:27 localhost podman[219189]: 2025-11-27 09:32:27.660944358 +0000 UTC m=+0.171617224 container start d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd) Nov 27 04:32:27 localhost podman[219189]: multipathd Nov 27 04:32:27 localhost systemd[1]: Started multipathd container. Nov 27 04:32:27 localhost multipathd[219203]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:32:27 localhost multipathd[219203]: INFO:__main__:Validating config file Nov 27 04:32:27 localhost multipathd[219203]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:32:27 localhost multipathd[219203]: INFO:__main__:Writing out command to execute Nov 27 04:32:27 localhost multipathd[219203]: ++ cat /run_command Nov 27 04:32:27 localhost multipathd[219203]: + CMD='/usr/sbin/multipathd -d' Nov 27 04:32:27 localhost multipathd[219203]: + ARGS= Nov 27 04:32:27 localhost multipathd[219203]: + sudo kolla_copy_cacerts Nov 27 04:32:27 localhost podman[219211]: 2025-11-27 09:32:27.743713213 +0000 UTC m=+0.078221797 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=starting, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:32:27 localhost multipathd[219203]: + [[ ! -n '' ]] Nov 27 04:32:27 localhost multipathd[219203]: + . kolla_extend_start Nov 27 04:32:27 localhost multipathd[219203]: Running command: '/usr/sbin/multipathd -d' Nov 27 04:32:27 localhost multipathd[219203]: + echo 'Running command: '\''/usr/sbin/multipathd -d'\''' Nov 27 04:32:27 localhost multipathd[219203]: + umask 0022 Nov 27 04:32:27 localhost multipathd[219203]: + exec /usr/sbin/multipathd -d Nov 27 04:32:27 localhost podman[219211]: 2025-11-27 09:32:27.753428938 +0000 UTC m=+0.087937502 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:32:27 localhost multipathd[219203]: 10177.943612 | --------start up-------- Nov 27 04:32:27 localhost multipathd[219203]: 10177.943633 | read /etc/multipath.conf Nov 27 04:32:27 localhost multipathd[219203]: 10177.947120 | path checkers start up Nov 27 04:32:27 localhost podman[219211]: unhealthy Nov 27 04:32:27 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:32:27 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Failed with result 'exit-code'. Nov 27 04:32:29 localhost python3.9[219351]: ansible-ansible.builtin.file Invoked with path=/etc/multipath/.multipath_restart_required state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25653 DF PROTO=TCP SPT=34830 DPT=9101 SEQ=3648322424 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292939A0000000001030307) Nov 27 04:32:30 localhost python3.9[219461]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/modules-load.d selevel=s0 setype=etc_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 04:32:31 localhost python3.9[219571]: ansible-community.general.modprobe Invoked with name=nvme-fabrics state=present params= persistent=disabled Nov 27 04:32:32 localhost python3.9[219689]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/nvme-fabrics.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:32:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20873 DF PROTO=TCP SPT=40738 DPT=9105 SEQ=574597365 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12929FDB0000000001030307) Nov 27 04:32:33 localhost python3.9[219777]: ansible-ansible.legacy.copy Invoked with dest=/etc/modules-load.d/nvme-fabrics.conf mode=0644 src=/home/zuul/.ansible/tmp/ansible-tmp-1764235951.9146922-1832-109154506292606/.source.conf follow=False _original_basename=module-load.conf.j2 checksum=783c778f0c68cc414f35486f234cbb1cf3f9bbff backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:34 localhost python3.9[219887]: ansible-ansible.builtin.lineinfile Invoked with create=True dest=/etc/modules line=nvme-fabrics mode=0644 state=present path=/etc/modules encoding=utf-8 backrefs=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:35 localhost python3.9[219997]: ansible-ansible.builtin.systemd Invoked with name=systemd-modules-load.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:32:35 localhost systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 27 04:32:35 localhost systemd[1]: Stopped Load Kernel Modules. Nov 27 04:32:35 localhost systemd[1]: Stopping Load Kernel Modules... Nov 27 04:32:35 localhost systemd[1]: Starting Load Kernel Modules... Nov 27 04:32:35 localhost systemd-modules-load[220001]: Module 'msr' is built in Nov 27 04:32:35 localhost systemd[1]: Finished Load Kernel Modules. Nov 27 04:32:36 localhost python3.9[220111]: ansible-ansible.legacy.dnf Invoked with name=['nvme-cli'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:32:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25654 DF PROTO=TCP SPT=34830 DPT=9101 SEQ=3648322424 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292B3DA0000000001030307) Nov 27 04:32:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16492 DF PROTO=TCP SPT=49762 DPT=9882 SEQ=3446647008 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292B6B70000000001030307) Nov 27 04:32:40 localhost systemd[1]: Reloading. Nov 27 04:32:40 localhost systemd-sysv-generator[220148]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:40 localhost systemd-rc-local-generator[220143]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: Reloading. Nov 27 04:32:40 localhost systemd-sysv-generator[220188]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:40 localhost systemd-rc-local-generator[220184]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:40 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd-logind[761]: Watching system buttons on /dev/input/event0 (Power Button) Nov 27 04:32:41 localhost systemd-logind[761]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Nov 27 04:32:41 localhost lvm[220232]: PV /dev/loop3 online, VG ceph_vg0 is complete. Nov 27 04:32:41 localhost lvm[220233]: PV /dev/loop4 online, VG ceph_vg1 is complete. Nov 27 04:32:41 localhost lvm[220233]: VG ceph_vg1 finished Nov 27 04:32:41 localhost lvm[220232]: VG ceph_vg0 finished Nov 27 04:32:41 localhost systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. Nov 27 04:32:41 localhost systemd[1]: Starting man-db-cache-update.service... Nov 27 04:32:41 localhost systemd[1]: Reloading. Nov 27 04:32:41 localhost systemd-rc-local-generator[220278]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:41 localhost systemd-sysv-generator[220281]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:41 localhost systemd[1]: Queuing reload/restart jobs for marked units… Nov 27 04:32:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16494 DF PROTO=TCP SPT=49762 DPT=9882 SEQ=3446647008 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292C2DA0000000001030307) Nov 27 04:32:42 localhost systemd[1]: man-db-cache-update.service: Deactivated successfully. Nov 27 04:32:42 localhost systemd[1]: Finished man-db-cache-update.service. Nov 27 04:32:42 localhost systemd[1]: man-db-cache-update.service: Consumed 1.308s CPU time. Nov 27 04:32:42 localhost systemd[1]: run-r48b50c0e7d7b4f6cb95f7944e81b9eac.service: Deactivated successfully. Nov 27 04:32:43 localhost python3.9[221527]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:32:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:32:43.538 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:32:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:32:43.539 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:32:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:32:43.541 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:32:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:32:44 localhost podman[221549]: 2025-11-27 09:32:44.00233935 +0000 UTC m=+0.084236508 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:32:44 localhost podman[221549]: 2025-11-27 09:32:44.045056794 +0000 UTC m=+0.126953912 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:32:44 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:32:44 localhost python3.9[221665]: ansible-ansible.builtin.file Invoked with mode=0644 path=/etc/ssh/ssh_known_hosts state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:32:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59700 DF PROTO=TCP SPT=33182 DPT=9102 SEQ=2321754763 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292D1490000000001030307) Nov 27 04:32:45 localhost python3.9[221775]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:32:45 localhost systemd[1]: Reloading. Nov 27 04:32:45 localhost systemd-rc-local-generator[221803]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:32:45 localhost systemd-sysv-generator[221806]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:46 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:32:47 localhost python3.9[221919]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:32:47 localhost network[221936]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:32:47 localhost network[221937]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:32:47 localhost network[221938]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:32:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47528 DF PROTO=TCP SPT=35396 DPT=9100 SEQ=2180046574 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292D95A0000000001030307) Nov 27 04:32:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:32:50 localhost podman[221973]: 2025-11-27 09:32:50.986648829 +0000 UTC m=+0.081485759 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:32:51 localhost podman[221973]: 2025-11-27 09:32:51.021868063 +0000 UTC m=+0.116704953 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125) Nov 27 04:32:51 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:32:51 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:32:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47529 DF PROTO=TCP SPT=35396 DPT=9100 SEQ=2180046574 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292E91A0000000001030307) Nov 27 04:32:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58595 DF PROTO=TCP SPT=35632 DPT=9101 SEQ=1980940352 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292F11A0000000001030307) Nov 27 04:32:56 localhost python3.9[222189]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_compute.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:56 localhost python3.9[222300]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_migration_target.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25341 DF PROTO=TCP SPT=33542 DPT=9105 SEQ=2534243965 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1292FD5A0000000001030307) Nov 27 04:32:57 localhost python3.9[222411]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_api_cron.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:32:58 localhost podman[222502]: 2025-11-27 09:32:58.007011077 +0000 UTC m=+0.090699139 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=starting, managed_by=edpm_ansible, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:32:58 localhost podman[222502]: 2025-11-27 09:32:58.018356577 +0000 UTC m=+0.102044619 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.build-date=20251125, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:32:58 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:32:58 localhost python3.9[222533]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_api.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:59 localhost python3.9[222653]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_conductor.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:32:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58597 DF PROTO=TCP SPT=35632 DPT=9101 SEQ=1980940352 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129308DB0000000001030307) Nov 27 04:33:00 localhost python3.9[222764]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_metadata.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:33:01 localhost python3.9[222875]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_scheduler.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:33:02 localhost python3.9[222986]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_vnc_proxy.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:33:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25343 DF PROTO=TCP SPT=33542 DPT=9105 SEQ=2534243965 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293151A0000000001030307) Nov 27 04:33:04 localhost python3.9[223097]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:04 localhost python3.9[223207]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:05 localhost python3.9[223317]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_api_cron.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:06 localhost python3.9[223427]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_api.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:07 localhost python3.9[223537]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_conductor.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:07 localhost python3.9[223647]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_metadata.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=58598 DF PROTO=TCP SPT=35632 DPT=9101 SEQ=1980940352 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129329DA0000000001030307) Nov 27 04:33:08 localhost python3.9[223757]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_scheduler.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28376 DF PROTO=TCP SPT=57476 DPT=9882 SEQ=2580789555 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12932BE60000000001030307) Nov 27 04:33:08 localhost python3.9[223867]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_vnc_proxy.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:09 localhost python3.9[223977]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:10 localhost python3.9[224087]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:10 localhost python3.9[224197]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_api_cron.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:11 localhost python3.9[224307]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_api.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28378 DF PROTO=TCP SPT=57476 DPT=9882 SEQ=2580789555 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129337DA0000000001030307) Nov 27 04:33:11 localhost python3.9[224417]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_conductor.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:12 localhost python3.9[224527]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_metadata.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:13 localhost python3.9[224637]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_scheduler.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:14 localhost python3.9[224747]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_vnc_proxy.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:33:14 localhost systemd[1]: tmp-crun.pLt2Ha.mount: Deactivated successfully. Nov 27 04:33:14 localhost podman[224765]: 2025-11-27 09:33:14.984755727 +0000 UTC m=+0.085286948 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_id=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:33:15 localhost podman[224765]: 2025-11-27 09:33:15.056051567 +0000 UTC m=+0.156582818 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:33:15 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:33:15 localhost python3.9[224883]: ansible-ansible.legacy.command Invoked with _raw_params=if systemctl is-active certmonger.service; then#012 systemctl disable --now certmonger.service#012 test -f /etc/systemd/system/certmonger.service || systemctl mask certmonger.service#012fi#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1789 DF PROTO=TCP SPT=49564 DPT=9102 SEQ=1865255875 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129346790000000001030307) Nov 27 04:33:16 localhost python3.9[224993]: ansible-ansible.builtin.find Invoked with file_type=any hidden=True paths=['/var/lib/certmonger/requests'] patterns=[] read_whole_file=False age_stamp=mtime recurse=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:33:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28553 DF PROTO=TCP SPT=38292 DPT=9100 SEQ=2634408835 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12934E9A0000000001030307) Nov 27 04:33:17 localhost python3.9[225103]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:33:17 localhost systemd[1]: Reloading. Nov 27 04:33:17 localhost systemd-rc-local-generator[225125]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:33:17 localhost systemd-sysv-generator[225129]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:17 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:33:18 localhost python3.9[225249]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_compute.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:19 localhost python3.9[225360]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_migration_target.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:19 localhost python3.9[225471]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_api_cron.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:20 localhost python3.9[225618]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_api.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:33:21 localhost podman[225793]: 2025-11-27 09:33:21.189002802 +0000 UTC m=+0.104895021 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 04:33:21 localhost podman[225809]: 2025-11-27 09:33:21.241102361 +0000 UTC m=+0.098503200 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, vcs-type=git, io.openshift.expose-services=, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, version=7, GIT_CLEAN=True, ceph=True, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., RELEASE=main, GIT_BRANCH=main, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph) Nov 27 04:33:21 localhost podman[225793]: 2025-11-27 09:33:21.272980801 +0000 UTC m=+0.188872950 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:33:21 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:33:21 localhost python3.9[225792]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_conductor.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:21 localhost podman[225809]: 2025-11-27 09:33:21.340153176 +0000 UTC m=+0.197553985 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, architecture=x86_64, com.redhat.component=rhceph-container, name=rhceph, CEPH_POINT_RELEASE=, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, GIT_CLEAN=True, ceph=True, version=7, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, release=553, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., io.openshift.expose-services=, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux ) Nov 27 04:33:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28554 DF PROTO=TCP SPT=38292 DPT=9100 SEQ=2634408835 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12935E5A0000000001030307) Nov 27 04:33:22 localhost python3.9[226031]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_metadata.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:23 localhost python3.9[226175]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_scheduler.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8737 DF PROTO=TCP SPT=45280 DPT=9101 SEQ=3046831831 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293665A0000000001030307) Nov 27 04:33:23 localhost python3.9[226304]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_vnc_proxy.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:33:25 localhost python3.9[226415]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25656 DF PROTO=TCP SPT=34830 DPT=9101 SEQ=3648322424 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129371DA0000000001030307) Nov 27 04:33:27 localhost python3.9[226525]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/containers setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:27 localhost python3.9[226635]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/nova_nvme_cleaner setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:33:28 localhost podman[226746]: 2025-11-27 09:33:28.309208867 +0000 UTC m=+0.103455460 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=multipathd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 04:33:28 localhost podman[226746]: 2025-11-27 09:33:28.324976102 +0000 UTC m=+0.119222645 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:33:28 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:33:28 localhost python3.9[226745]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:29 localhost python3.9[226875]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/_nova_secontext setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28555 DF PROTO=TCP SPT=38292 DPT=9100 SEQ=2634408835 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12937DDA0000000001030307) Nov 27 04:33:30 localhost python3.9[226985]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/nova/instances setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:30 localhost python3.9[227095]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/etc/ceph setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:31 localhost python3.9[227205]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/etc/multipath setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:32 localhost python3.9[227315]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/etc/nvme setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:32 localhost python3.9[227425]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/run/openvswitch setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1890 DF PROTO=TCP SPT=53950 DPT=9105 SEQ=1044711163 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12938A1A0000000001030307) Nov 27 04:33:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=8740 DF PROTO=TCP SPT=45280 DPT=9101 SEQ=3046831831 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12939DDA0000000001030307) Nov 27 04:33:38 localhost python3.9[227535]: ansible-ansible.builtin.getent Invoked with database=passwd key=nova fail_key=True service=None split=None Nov 27 04:33:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37565 DF PROTO=TCP SPT=36362 DPT=9882 SEQ=1853156246 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293A1170000000001030307) Nov 27 04:33:39 localhost python3.9[227646]: ansible-ansible.builtin.group Invoked with gid=42436 name=nova state=present force=False system=False local=False non_unique=False gid_min=None gid_max=None Nov 27 04:33:41 localhost python3.9[227762]: ansible-ansible.builtin.user Invoked with comment=nova user group=nova groups=['libvirt'] name=nova shell=/bin/sh state=present uid=42436 non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.localdomain update_password=always home=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None password_expire_account_disable=None uid_min=None uid_max=None Nov 27 04:33:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=37567 DF PROTO=TCP SPT=36362 DPT=9882 SEQ=1853156246 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293AD1A0000000001030307) Nov 27 04:33:42 localhost sshd[227788]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:33:43 localhost systemd-logind[761]: New session 55 of user zuul. Nov 27 04:33:43 localhost systemd[1]: Started Session 55 of User zuul. Nov 27 04:33:43 localhost systemd[1]: session-55.scope: Deactivated successfully. Nov 27 04:33:43 localhost systemd-logind[761]: Session 55 logged out. Waiting for processes to exit. Nov 27 04:33:43 localhost systemd-logind[761]: Removed session 55. Nov 27 04:33:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:33:43.539 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:33:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:33:43.541 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:33:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:33:43.542 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:33:43 localhost python3.9[227899]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/config.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:44 localhost python3.9[227985]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/config.json mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236023.399884-3391-239819878260970/.source.json follow=False _original_basename=config.json.j2 checksum=b51012bfb0ca26296dcf3793a2f284446fb1395e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:44 localhost python3.9[228093]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/nova-blank.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:45 localhost python3.9[228148]: ansible-ansible.legacy.file Invoked with mode=0644 setype=container_file_t dest=/var/lib/openstack/config/nova/nova-blank.conf _original_basename=nova-blank.conf recurse=False state=file path=/var/lib/openstack/config/nova/nova-blank.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=10883 DF PROTO=TCP SPT=40526 DPT=9102 SEQ=280567007 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293BBA90000000001030307) Nov 27 04:33:45 localhost python3.9[228256]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/ssh-config follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:33:45 localhost podman[228257]: 2025-11-27 09:33:45.992933839 +0000 UTC m=+0.086835941 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller) Nov 27 04:33:46 localhost podman[228257]: 2025-11-27 09:33:46.055968546 +0000 UTC m=+0.149870618 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:33:46 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:33:46 localhost python3.9[228367]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/ssh-config mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236025.444038-3391-219388362504088/.source follow=False _original_basename=ssh-config checksum=4297f735c41bdc1ff52d72e6f623a02242f37958 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:46 localhost python3.9[228475]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/02-nova-host-specific.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:47 localhost python3.9[228561]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/02-nova-host-specific.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236026.4508948-3391-213086395702129/.source.conf follow=False _original_basename=02-nova-host-specific.conf.j2 checksum=ed20ceeae546fb32715585d43ee52c52842c4309 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13481 DF PROTO=TCP SPT=40110 DPT=9100 SEQ=815931384 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293C3DA0000000001030307) Nov 27 04:33:47 localhost python3.9[228669]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/nova_statedir_ownership.py follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:48 localhost python3.9[228755]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/nova_statedir_ownership.py mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236027.4850795-3391-49357463215999/.source.py follow=False _original_basename=nova_statedir_ownership.py checksum=c6c8a3cfefa5efd60ceb1408c4e977becedb71e2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:49 localhost python3.9[228863]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/run-on-host follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:49 localhost python3.9[228949]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/run-on-host mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236028.545499-3391-126602910898945/.source follow=False _original_basename=run-on-host checksum=93aba8edc83d5878604a66d37fea2f12b60bdea2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:50 localhost python3.9[229059]: ansible-ansible.builtin.file Invoked with group=nova mode=0700 owner=nova path=/home/nova/.ssh state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:51 localhost python3.9[229169]: ansible-ansible.legacy.copy Invoked with dest=/home/nova/.ssh/authorized_keys group=nova mode=0600 owner=nova remote_src=True src=/var/lib/openstack/config/nova/ssh-publickey backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:33:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=13482 DF PROTO=TCP SPT=40110 DPT=9100 SEQ=815931384 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293D39B0000000001030307) Nov 27 04:33:51 localhost systemd[1]: tmp-crun.v2sB1b.mount: Deactivated successfully. Nov 27 04:33:51 localhost podman[229280]: 2025-11-27 09:33:51.765790062 +0000 UTC m=+0.084835886 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:33:51 localhost podman[229280]: 2025-11-27 09:33:51.798883075 +0000 UTC m=+0.117928859 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:33:51 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:33:51 localhost python3.9[229279]: ansible-ansible.builtin.stat Invoked with path=/var/lib/nova/compute_id follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:33:52 localhost python3.9[229407]: ansible-ansible.builtin.file Invoked with group=nova mode=0400 owner=nova path=/var/lib/nova/compute_id state=file recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:33:53 localhost python3.9[229515]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:33:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53749 DF PROTO=TCP SPT=53328 DPT=9101 SEQ=3153789417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293DB9A0000000001030307) Nov 27 04:33:54 localhost python3.9[229625]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/containers/nova_compute.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:55 localhost python3.9[229711]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/containers/nova_compute.json mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236033.660431-3766-31894434560852/.source.json follow=False _original_basename=nova_compute.json.j2 checksum=211ffd0bca4b407eb4de45a749ef70116a7806fd backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:55 localhost python3.9[229819]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/containers/nova_compute_init.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:33:56 localhost python3.9[229905]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/containers/nova_compute_init.json mode=0700 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236035.5110557-3812-188020355783506/.source.json follow=False _original_basename=nova_compute_init.json.j2 checksum=60b024e6db49dc6e700fc0d50263944d98d4c034 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:33:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20086 DF PROTO=TCP SPT=33424 DPT=9105 SEQ=2422489915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293E79A0000000001030307) Nov 27 04:33:57 localhost python3.9[230015]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/containers config_pattern=nova_compute_init.json debug=False Nov 27 04:33:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:33:58 localhost systemd[1]: tmp-crun.cCgBSc.mount: Deactivated successfully. Nov 27 04:33:58 localhost podman[230126]: 2025-11-27 09:33:58.658168414 +0000 UTC m=+0.093789037 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, tcib_managed=true, managed_by=edpm_ansible, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:33:58 localhost podman[230126]: 2025-11-27 09:33:58.671747952 +0000 UTC m=+0.107368535 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, container_name=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:33:58 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:33:58 localhost python3.9[230125]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:33:59 localhost sshd[230185]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:33:59 localhost sshd[230253]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:33:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53751 DF PROTO=TCP SPT=53328 DPT=9101 SEQ=3153789417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293F35A0000000001030307) Nov 27 04:33:59 localhost python3[230255]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/containers config_id=edpm config_overrides={} config_patterns=nova_compute_init.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:34:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20088 DF PROTO=TCP SPT=33424 DPT=9105 SEQ=2422489915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1293FF5A0000000001030307) Nov 27 04:34:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53752 DF PROTO=TCP SPT=53328 DPT=9101 SEQ=3153789417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129413DB0000000001030307) Nov 27 04:34:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28656 DF PROTO=TCP SPT=54262 DPT=9882 SEQ=1897157412 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129416470000000001030307) Nov 27 04:34:10 localhost podman[230268]: 2025-11-27 09:33:59.97567844 +0000 UTC m=+0.047351102 image pull quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified Nov 27 04:34:10 localhost podman[230330]: Nov 27 04:34:10 localhost podman[230330]: 2025-11-27 09:34:10.345553192 +0000 UTC m=+0.084559718 container create 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=nova_compute_init, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=edpm, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}) Nov 27 04:34:10 localhost podman[230330]: 2025-11-27 09:34:10.305781623 +0000 UTC m=+0.044788189 image pull quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified Nov 27 04:34:10 localhost python3[230255]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name nova_compute_init --conmon-pidfile /run/nova_compute_init.pid --env NOVA_STATEDIR_OWNERSHIP_SKIP=/var/lib/nova/compute_id --env __OS_DEBUG=False --label config_id=edpm --label container_name=nova_compute_init --label managed_by=edpm_ansible --label config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']} --log-driver journald --log-level info --network none --privileged=False --security-opt label=disable --user root --volume /dev/log:/dev/log --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z --volume /var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init Nov 27 04:34:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=28658 DF PROTO=TCP SPT=54262 DPT=9882 SEQ=1897157412 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294225B0000000001030307) Nov 27 04:34:12 localhost python3.9[230477]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:34:13 localhost python3.9[230589]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/containers config_pattern=nova_compute.json debug=False Nov 27 04:34:14 localhost python3.9[230699]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:34:15 localhost python3[230809]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/containers config_id=edpm config_overrides={} config_patterns=nova_compute.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:34:15 localhost python3[230809]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "b65793e7266422f5b94c32d109b906c8ffd974cf2ddf0b6929e463e29e05864a",#012 "Digest": "sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:36:07.10279245Z",#012 "Config": {#012 "User": "nova",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 1211782527,#012 "VirtualSize": 1211782527,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/c3914bdda39f47c0c497a56396d11c84b489b87df2bfd019b00ddced1e1ae309/diff:/var/lib/containers/storage/overlay/f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:c136b33417f134a3b932677bcf7a2df089c29f20eca250129eafd2132d4708bb",#012 "sha256:7913bde445307e7f24767d9149b2e7f498930793ac9f073ccec69b608c009d31",#012 "sha256:084b2323a717fe711217b0ec21da61f4804f7a0d506adae935888421b80809cf"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "nova",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 {#012 Nov 27 04:34:15 localhost podman[230861]: 2025-11-27 09:34:15.505571699 +0000 UTC m=+0.089011284 container remove ed18f348a0e7c3001cf1ab94781267bbff58bac7a1fd530269f8dc60a3daa409 (image=registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1, name=nova_compute, config_id=tripleo_step5, release=1761123044, vcs-ref=d13aeaae6d02e9d9273775f1920879be7af2cf2d, io.openshift.expose-services=, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'LIBGUESTFS_BACKEND': 'direct', 'TRIPLEO_CONFIG_HASH': '2c87de3317f94758d1bec36af3e86047-4018231ef757a0bfd34181dc6dc2933a'}, 'healthcheck': {'test': '/openstack/healthcheck 5672'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-nova-compute:17.1', 'ipc': 'host', 'net': 'host', 'privileged': True, 'restart': 'always', 'start_order': 3, 'ulimit': ['nofile=131072', 'memlock=67108864'], 'user': 'nova', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/log/containers/nova:/var/log/nova', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro', '/var/lib/kolla/config_files/nova_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/nova_libvirt:/var/lib/kolla/config_files/src:ro', '/var/lib/config-data/puppet-generated/iscsid/etc/iscsi:/var/lib/kolla/config_files/src-iscsid:ro', '/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:z', '/dev:/dev', '/lib/modules:/lib/modules:ro', '/run:/run', '/run/nova:/run/nova:z', '/var/lib/iscsi:/var/lib/iscsi:z', '/var/lib/libvirt:/var/lib/libvirt:shared', '/sys/class/net:/sys/class/net', '/sys/bus/pci:/sys/bus/pci', '/boot:/boot:ro', '/var/lib/nova:/var/lib/nova:shared']}, io.k8s.display-name=Red Hat OpenStack Platform 17.1 nova-compute, summary=Red Hat OpenStack Platform 17.1 nova-compute, org.opencontainers.image.revision=d13aeaae6d02e9d9273775f1920879be7af2cf2d, version=17.1.12, name=rhosp17/openstack-nova-compute, com.redhat.component=openstack-nova-compute-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, tcib_managed=true, managed_by=tripleo_ansible, batch=17.1_20251118.1, build-date=2025-11-19T00:36:58Z, architecture=x86_64, vcs-type=git, io.k8s.description=Red Hat OpenStack Platform 17.1 nova-compute, container_name=nova_compute, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-nova-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, io.buildah.version=1.41.4, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 nova-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, distribution-scope=public, url=https://www.redhat.com, maintainer=OpenStack TripleO Team, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream) Nov 27 04:34:15 localhost python3[230809]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman rm --force nova_compute Nov 27 04:34:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=15298 DF PROTO=TCP SPT=44928 DPT=9102 SEQ=2922285301 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129430D90000000001030307) Nov 27 04:34:15 localhost podman[230875]: Nov 27 04:34:15 localhost podman[230875]: 2025-11-27 09:34:15.615035831 +0000 UTC m=+0.089882017 container create 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, org.label-schema.build-date=20251125, container_name=nova_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:34:15 localhost podman[230875]: 2025-11-27 09:34:15.571035553 +0000 UTC m=+0.045881759 image pull quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified Nov 27 04:34:15 localhost python3[230809]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name nova_compute --conmon-pidfile /run/nova_compute.pid --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --label config_id=edpm --label container_name=nova_compute --label managed_by=edpm_ansible --label config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']} --log-driver journald --log-level info --network host --pid host --privileged=True --user nova --volume /var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro --volume /var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z --volume /etc/localtime:/etc/localtime:ro --volume /lib/modules:/lib/modules:ro --volume /dev:/dev --volume /var/lib/libvirt:/var/lib/libvirt --volume /run/libvirt:/run/libvirt:shared --volume /var/lib/nova:/var/lib/nova:shared --volume /var/lib/iscsi:/var/lib/iscsi --volume /etc/multipath:/etc/multipath:z --volume /etc/multipath.conf:/etc/multipath.conf:ro --volume /etc/iscsi:/etc/iscsi:ro --volume /etc/nvme:/etc/nvme --volume /var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro --volume /etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified kolla_start Nov 27 04:34:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:34:16 localhost podman[231022]: 2025-11-27 09:34:16.45674942 +0000 UTC m=+0.072334458 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, container_name=ovn_controller, org.label-schema.build-date=20251125, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:34:16 localhost podman[231022]: 2025-11-27 09:34:16.488832935 +0000 UTC m=+0.104417963 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=ovn_controller, managed_by=edpm_ansible, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0) Nov 27 04:34:16 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:34:16 localhost python3.9[231021]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:34:17 localhost python3.9[231157]: ansible-file Invoked with path=/etc/systemd/system/edpm_nova_compute.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:34:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2449 DF PROTO=TCP SPT=39384 DPT=9100 SEQ=3975749257 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294391A0000000001030307) Nov 27 04:34:17 localhost python3.9[231266]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236057.4454765-4086-127203224769509/source dest=/etc/systemd/system/edpm_nova_compute.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:34:18 localhost python3.9[231321]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:34:18 localhost systemd[1]: Reloading. Nov 27 04:34:18 localhost systemd-rc-local-generator[231344]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:34:18 localhost systemd-sysv-generator[231348]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:18 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost python3.9[231411]: ansible-systemd Invoked with state=restarted name=edpm_nova_compute.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:34:19 localhost systemd[1]: Reloading. Nov 27 04:34:19 localhost systemd-rc-local-generator[231435]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:34:19 localhost systemd-sysv-generator[231441]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:19 localhost systemd[1]: Starting nova_compute container... Nov 27 04:34:20 localhost systemd[1]: Started libcrun container. Nov 27 04:34:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/nvme supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/multipath supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:20 localhost podman[231451]: 2025-11-27 09:34:20.079544005 +0000 UTC m=+0.120408618 container init 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=edpm, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, maintainer=OpenStack Kubernetes Operator team, container_name=nova_compute, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:34:20 localhost podman[231451]: 2025-11-27 09:34:20.088867385 +0000 UTC m=+0.129731998 container start 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, container_name=nova_compute, io.buildah.version=1.41.3, config_id=edpm, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:34:20 localhost podman[231451]: nova_compute Nov 27 04:34:20 localhost nova_compute[231466]: + sudo -E kolla_set_configs Nov 27 04:34:20 localhost systemd[1]: Started nova_compute container. Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Validating config file Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying service configuration files Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Deleting /etc/nova/nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/01-nova.conf to /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/03-ceph-nova.conf to /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/99-nova-compute-cells-workarounds.conf to /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/02-nova-host-specific.conf to /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Deleting /etc/ceph Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Creating directory /etc/ceph Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/ceph Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.conf to /etc/ceph/ceph.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.client.openstack.keyring to /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-privatekey to /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Deleting /var/lib/nova/.ssh/config Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-config to /var/lib/nova/.ssh/config Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Deleting /usr/sbin/iscsiadm Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Copying /var/lib/kolla/config_files/run-on-host to /usr/sbin/iscsiadm Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /usr/sbin/iscsiadm Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Writing out command to execute Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:20 localhost nova_compute[231466]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:34:20 localhost nova_compute[231466]: ++ cat /run_command Nov 27 04:34:20 localhost nova_compute[231466]: + CMD=nova-compute Nov 27 04:34:20 localhost nova_compute[231466]: + ARGS= Nov 27 04:34:20 localhost nova_compute[231466]: + sudo kolla_copy_cacerts Nov 27 04:34:20 localhost nova_compute[231466]: + [[ ! -n '' ]] Nov 27 04:34:20 localhost nova_compute[231466]: + . kolla_extend_start Nov 27 04:34:20 localhost nova_compute[231466]: Running command: 'nova-compute' Nov 27 04:34:20 localhost nova_compute[231466]: + echo 'Running command: '\''nova-compute'\''' Nov 27 04:34:20 localhost nova_compute[231466]: + umask 0022 Nov 27 04:34:20 localhost nova_compute[231466]: + exec nova-compute Nov 27 04:34:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=2450 DF PROTO=TCP SPT=39384 DPT=9100 SEQ=3975749257 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129448DA0000000001030307) Nov 27 04:34:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:34:21 localhost nova_compute[231466]: 2025-11-27 09:34:21.971 231470 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:21 localhost nova_compute[231466]: 2025-11-27 09:34:21.972 231470 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:21 localhost nova_compute[231466]: 2025-11-27 09:34:21.972 231470 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:21 localhost nova_compute[231466]: 2025-11-27 09:34:21.972 231470 INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs#033[00m Nov 27 04:34:21 localhost systemd[1]: tmp-crun.77YK5g.mount: Deactivated successfully. Nov 27 04:34:21 localhost podman[231498]: 2025-11-27 09:34:21.990059926 +0000 UTC m=+0.087769499 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:34:22 localhost podman[231498]: 2025-11-27 09:34:22.020397761 +0000 UTC m=+0.118107294 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:34:22 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.097 231470 DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.118 231470 DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 1 in 0.021s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.119 231470 DEBUG oslo_concurrency.processutils [-] 'grep -F node.session.scan /sbin/iscsiadm' failed. Not Retrying. execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:473#033[00m Nov 27 04:34:22 localhost python3.9[231635]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner_healthcheck.service follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.545 231470 INFO nova.virt.driver [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Loading compute driver 'libvirt.LibvirtDriver'#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.673 231470 INFO nova.compute.provider_config [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access.#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.682 231470 WARNING nova.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Current Nova version does not support computes older than Yoga but the minimum compute service level in your cell is 57 and the oldest supported service level is 61.: nova.exception.TooOldComputeService: Current Nova version does not support computes older than Yoga but the minimum compute service level in your cell is 57 and the oldest supported service level is 61.#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.682 231470 DEBUG oslo_concurrency.lockutils [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_concurrency.lockutils [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_concurrency.lockutils [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Full set of CONF: _wait_for_exit_or_signal /usr/lib/python3.9/site-packages/oslo_service/service.py:362#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.683 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] command line args: [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] allow_resize_to_same_host = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] arq_binding_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] backdoor_port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] backdoor_socket = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] block_device_allocate_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.684 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] block_device_allocate_retries_interval = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cert = self.pem log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute_driver = libvirt.LibvirtDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute_monitors = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] config_dir = ['/etc/nova/nova.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] config_drive_format = iso9660 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.685 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] console_host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] control_exchange = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cpu_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] default_access_ip_network_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.686 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] default_availability_zone = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] default_ephemeral_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] default_schedule_zone = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] disk_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] enable_new_services = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] enabled_apis = ['osapi_compute', 'metadata'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.687 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] enabled_ssl_apis = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] flat_injected = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] force_config_drive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] force_raw_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] heal_instance_info_cache_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.688 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] initial_cpu_allocation_ratio = 4.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] initial_disk_allocation_ratio = 0.9 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] initial_ram_allocation_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] injected_network_template = /usr/lib/python3.9/site-packages/nova/virt/interfaces.template log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_build_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.689 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_delete_interval = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_name_template = instance-%08x log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_usage_audit = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_usage_audit_period = month log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.690 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] instances_path = /var/lib/nova/instances log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.691 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] internal_service_availability_zone = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.691 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.691 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] live_migration_retry_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.691 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.691 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] log_rotation_type = size log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.692 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] long_rpc_timeout = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.693 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_concurrent_builds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_concurrent_live_migrations = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_concurrent_snapshots = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_local_block_devices = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_logfile_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] max_logfile_size_mb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.694 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] maximum_instance_delete_attempts = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metadata_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metadata_listen_port = 8775 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metadata_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] migrate_max_retries = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] mkisofs_cmd = /usr/bin/mkisofs log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.695 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] my_block_storage_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] my_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] network_allocate_retries = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] osapi_compute_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] osapi_compute_listen_port = 8774 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.696 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] osapi_compute_unique_server_name_scope = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] osapi_compute_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] password_length = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] periodic_enable = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] periodic_fuzzy_delay = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] pointer_model = usbtablet log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] preallocate_images = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.697 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] pybasedir = /usr/lib/python3.9/site-packages log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ram_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reboot_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.698 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reclaim_instance_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] record = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reimage_timeout_per_gb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] report_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rescue_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reserved_host_cpus = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reserved_host_disk_mb = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.699 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reserved_host_memory_mb = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] reserved_huge_pages = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] resize_confirm_window = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] resize_fs_using_block_device = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] resume_guests_state_on_host_boot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rootwrap_config = /etc/nova/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rpc_response_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] run_external_periodic_tasks = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.700 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] running_deleted_instance_action = reap log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] running_deleted_instance_poll_interval = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] running_deleted_instance_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler_instance_sync_interval = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_down_time = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] servicegroup_driver = db log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] shelved_offload_time = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] shelved_poll_interval = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.701 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] source_is_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ssl_only = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] state_path = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] sync_power_state_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] sync_power_state_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] tempdir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.702 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] timeout_nbd = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] update_resources_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_cow_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.703 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_rootwrap_daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vcpu_pin_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plugging_is_fatal = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plugging_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] virt_mkfs = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] volume_usage_poll_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.704 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] web = /usr/share/spice-html5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_concurrency.disable_process_locking = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_concurrency.lock_path = /var/lib/nova/tmp log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_metrics.metrics_buffer_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_metrics.metrics_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_metrics.metrics_process_name = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.705 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.auth_strategy = keystone log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.compute_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.dhcp_domain = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.enable_instance_password = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.706 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.glance_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.instance_list_cells_batch_fixed_size = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.instance_list_cells_batch_strategy = distributed log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.instance_list_per_project_cells = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.list_records_by_skipping_down_cells = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.local_metadata_per_cell = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.max_limit = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.707 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.metadata_cache_expiration = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.neutron_default_tenant_id = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.use_forwarded_for = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.use_neutron_default_nets = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_dynamic_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_dynamic_failure_fatal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_dynamic_read_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_dynamic_ssl_certfile = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.708 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_dynamic_targets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_jsonfile_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api.vendordata_providers = ['StaticJSON'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.backend = oslo_cache.dict log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.backend_argument = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.config_prefix = cache.oslo log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.dead_timeout = 60.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.709 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.debug_cache_backend = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.enable_retry_client = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.enable_socket_keepalive = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.expiration_time = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.hashclient_retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.hashclient_retry_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_dead_retry = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.710 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_password = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_pool_connection_get_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_pool_flush_on_reconnect = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_pool_maxsize = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_pool_unused_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_sasl_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_servers = ['localhost:11211'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.711 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_socket_timeout = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.memcache_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.proxies = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.retry_delay = 0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.socket_keepalive_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.socket_keepalive_idle = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.socket_keepalive_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.712 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.tls_allowed_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.tls_cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.tls_certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.tls_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cache.tls_keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.713 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.catalog_info = volumev3:cinderv3:internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.cross_az_attach = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.endpoint_template = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.714 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.os_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cinder.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.consecutive_build_service_disable_threshold = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.cpu_dedicated_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.cpu_shared_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.715 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.image_type_exclude_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.live_migration_wait_for_vif_plug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.max_concurrent_disk_ops = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.max_disk_devices_to_attach = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.packing_host_numa_cells_allocation_strategy = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.provider_config_location = /etc/nova/provider_config/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.resource_provider_association_refresh = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.716 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.shutdown_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] conductor.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] console.allowed_origins = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] console.ssl_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] console.ssl_minimum_version = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] consoleauth.token_ttl = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.717 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.718 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.service_type = accelerator log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.719 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] cyborg.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.720 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.721 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.722 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.723 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.724 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] api_database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] devices.enabled_mdev_types = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.725 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ephemeral_storage_encryption.cipher = aes-xts-plain64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ephemeral_storage_encryption.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ephemeral_storage_encryption.key_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.api_servers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.726 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.default_trusted_certificate_ids = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.enable_certificate_validation = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.enable_rbd_download = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.727 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.num_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.rbd_ceph_conf = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.rbd_pool = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.rbd_user = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.728 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.service_type = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.729 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.verify_glance_signatures = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] glance.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] guestfs.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.config_drive_cdrom = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.config_drive_inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.dynamic_memory_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.enable_instance_metrics_collection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.enable_remotefx = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.730 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.instances_path_share = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.iscsi_initiator_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.limit_cpu_features = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.mounted_disk_query_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.mounted_disk_query_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.power_state_check_timeframe = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.power_state_event_polling_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.731 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.qemu_img_cmd = qemu-img.exe log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.use_multipath_io = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.volume_attach_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.volume_attach_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.vswitch_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] hyperv.wait_soft_reboot_seconds = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] mks.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.732 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] mks.mksproxy_base_url = http://127.0.0.1:6090/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.manager_interval = 2400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.precache_concurrency = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.remove_unused_base_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.remove_unused_original_minimum_age_seconds = 86400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.remove_unused_resized_minimum_age_seconds = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] image_cache.subdirectory_name = _base log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.733 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.api_max_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.api_retry_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.734 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.partition_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.735 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.peer_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.serial_console_state_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.service_type = baremetal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.736 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ironic.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] key_manager.backend = barbican log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] key_manager.fixed_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.auth_endpoint = http://localhost/identity/v3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.barbican_api_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.737 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.barbican_endpoint = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.barbican_endpoint_type = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.barbican_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.738 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.number_of_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.retry_delay = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.send_service_user_token = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.verify_ssl = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican.verify_ssl_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.739 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] barbican_service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.740 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.approle_role_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.approle_secret_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.741 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.kv_mountpoint = secret log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.kv_version = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.root_token_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.use_ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.742 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vault.vault_url = http://127.0.0.1:8200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.743 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.service_type = identity log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.744 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] keystone.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.connection_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_mode = host-model log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_model_extra_flags = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.745 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_models = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_power_governor_high = performance log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_power_governor_low = powersave log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_power_management = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.cpu_power_management_strategy = cpu_state log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.device_detach_attempts = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.device_detach_timeout = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.disk_cachemodes = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.746 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.disk_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.enabled_perf_events = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.file_backed_memory = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.gid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.hw_disk_discard = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.hw_machine_type = ['x86_64=q35'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_rbd_ceph_conf = /etc/ceph/ceph.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.747 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_rbd_glance_copy_poll_interval = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_rbd_glance_copy_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_rbd_glance_store_name = default_backend log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_rbd_pool = vms log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_type = rbd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.images_volume_group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.inject_key = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.inject_partition = -2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.748 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.iscsi_iface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.iser_use_multipath = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_bandwidth = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_completion_timeout = 800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_downtime = 500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_downtime_delay = 75 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.749 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_downtime_steps = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_inbound_addr = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_permit_auto_converge = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_permit_post_copy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_scheme = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_timeout_action = force_complete log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_tunnelled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.750 231470 WARNING oslo_config.cfg [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] Deprecated: Option "live_migration_uri" from group "libvirt" is deprecated for removal ( Nov 27 04:34:22 localhost nova_compute[231466]: live_migration_uri is deprecated for removal in favor of two other options that Nov 27 04:34:22 localhost nova_compute[231466]: allow to change live migration scheme and target URI: ``live_migration_scheme`` Nov 27 04:34:22 localhost nova_compute[231466]: and ``live_migration_inbound_addr`` respectively. Nov 27 04:34:22 localhost nova_compute[231466]: ). Its value may be silently ignored in the future.#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_uri = qemu+ssh://nova@%s/system?keyfile=/var/lib/nova/.ssh/ssh-privatekey log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.live_migration_with_native_tls = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.max_queues = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.mem_stats_period_seconds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.nfs_mount_options = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.nfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.751 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_aoe_discover_tries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_iser_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_memory_encrypted_guests = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_nvme_discover_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_pcie_ports = 24 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.num_volume_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.pmem_namespaces = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.752 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.quobyte_client_cfg = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.quobyte_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rbd_destroy_volume_retries = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rbd_destroy_volume_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rbd_secret_uuid = e83f3b0c-4090-52df-95d4-ad9be8516692 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rbd_user = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.realtime_scheduler_priority = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.753 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.remote_filesystem_transport = ssh log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rescue_image_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rescue_kernel_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rescue_ramdisk_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rng_dev_path = /dev/urandom log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.rx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.smbfs_mount_options = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.754 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.smbfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.snapshot_compression = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.snapshot_image_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.snapshots_directory = /var/lib/nova/instances/snapshots log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.sparse_logical_volumes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.swtpm_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.swtpm_group = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.755 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.swtpm_user = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.sysinfo_serial = unique log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.tx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.uid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.use_virtio_for_bridges = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.virt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.volume_clear = zero log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.volume_clear_size = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.756 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.volume_use_multipath = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_cache_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_mount_group = qemu log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_mount_opts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_mount_perms = 0770 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.757 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.vzstorage_mount_user = stack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] libvirt.wait_soft_reboot_seconds = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.758 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.default_floating_pool = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.extension_sync_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.759 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.metadata_proxy_shared_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.ovs_bridge = br-int log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.physnets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.service_metadata_proxy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.760 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.service_type = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] neutron.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] notifications.bdms_in_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.761 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] notifications.default_level = INFO log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] notifications.notification_format = unversioned log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] notifications.notify_on_state_change = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] notifications.versioned_notifications_topics = ['versioned_notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] pci.alias = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] pci.device_spec = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] pci.report_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.762 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.763 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.764 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.project_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.project_name = service log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.765 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.service_type = placement log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.system_scope = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.766 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] placement.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.cores = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.count_usage_from_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.767 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.driver = nova.quota.DbQuotaDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.injected_file_content_bytes = 10240 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.injected_file_path_length = 255 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.injected_files = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.instances = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.key_pairs = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.metadata_items = 128 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.ram = 51200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.768 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.recheck_quota = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.server_group_members = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] quota.server_groups = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rdp.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] rdp.html5_proxy_base_url = http://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.discover_hosts_in_cells_interval = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.enable_isolated_aggregate_filtering = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.769 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.image_metadata_prefilter = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.limit_tenants_to_placement_aggregate = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.max_attempts = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.max_placement_results = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.placement_aggregate_required_for_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.query_placement_for_availability_zone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.query_placement_for_image_type_support = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.query_placement_for_routed_network_aggregates = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.770 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] scheduler.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.aggregate_image_properties_isolation_namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.aggregate_image_properties_isolation_separator = . log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.build_failure_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.cpu_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.771 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.disk_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.host_subset_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.image_properties_default_architecture = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.io_ops_weight_multiplier = -1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.isolated_hosts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.isolated_images = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.max_instances_per_host = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.772 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.max_io_ops_per_host = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.pci_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.pci_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.ram_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.shuffle_best_same_weighed_hosts = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.soft_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.773 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.track_instance_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metrics.required = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metrics.weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metrics.weight_of_unavailable = -10000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] metrics.weight_setting = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.774 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.base_url = ws://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.port_range = 10000:20000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.serialproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] serial_console.serialproxy_port = 6083 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.775 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.send_service_user_token = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.776 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.agent_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.html5proxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.html5proxy_port = 6082 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.image_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.777 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.jpeg_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.playback_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.server_listen = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.server_proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.streaming_mode = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] spice.zlib_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] upgrade_levels.baseapi = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.778 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] upgrade_levels.cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] upgrade_levels.compute = auto log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] upgrade_levels.conductor = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] upgrade_levels.scheduler = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.779 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vendordata_dynamic_auth.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.780 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.cache_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.cluster_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.connection_pool_size = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.console_delay_seconds = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.datastore_regex = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.host_ip = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.781 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.host_username = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.integration_bridge = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.maximum_objects = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.pbm_default_policy = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.pbm_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.pbm_wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.782 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.serial_log_dir = /opt/vmware/vspc log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.serial_port_proxy_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.serial_port_service_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.use_linked_clone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.vnc_keymap = en-us log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.vnc_port = 5900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vmware.vnc_port_total = 10000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.783 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.auth_schemes = ['none'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.novncproxy_base_url = http://nova-novncproxy-cell1-public-openstack.apps-crc.testing/vnc_lite.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.novncproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.novncproxy_port = 6080 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.server_listen = ::0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.784 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.server_proxyclient_address = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.vencrypt_ca_certs = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.vencrypt_client_cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vnc.vencrypt_client_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_compute_service_check_for_ffu = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_deep_image_inspection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_fallback_pcpu_query = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.785 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_group_policy_check_upcall = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_libvirt_livesnapshot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.disable_rootwrap = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.enable_numa_live_migration = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.enable_qemu_monitor_announce_self = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.handle_virt_lifecycle_events = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.libvirt_disable_apic = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.786 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.never_download_image_if_on_rbd = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.qemu_monitor_announce_self_count = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.qemu_monitor_announce_self_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.reserve_disk_resource_for_image_cache = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.skip_cpu_compare_at_startup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.skip_cpu_compare_on_dest = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.skip_hypervisor_version_check_on_lm = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.skip_reserve_in_use_ironic_nodes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.787 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.unified_limits_count_pcpu_as_vcpu = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.api_paste_config = api-paste.ini log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.client_socket_timeout = 900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.default_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.keep_alive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.max_header_line = 16384 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.788 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.secure_proxy_ssl_header = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.ssl_ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.ssl_cert_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.ssl_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.tcp_keepidle = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] zvm.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] zvm.cloud_connector_url = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.789 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] zvm.image_tmp_path = /var/lib/nova/images log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] zvm.reachable_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.enforce_new_defaults = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.enforce_scope = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.policy_default_rule = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.policy_dirs = ['policy.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.policy_file = policy.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.790 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.remote_content_type = application/x-www-form-urlencoded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.remote_ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.remote_ssl_client_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.remote_ssl_client_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_policy.remote_ssl_verify_server_crt = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_versionedobjects.fatal_exception_format_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.791 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] remote_debug.host = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] remote_debug.port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.amqp_durable_queues = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.792 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.793 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_quorum_queue = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.794 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.795 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_notifications.driver = ['noop'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.796 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.797 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.endpoint_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.798 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.project_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.project_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.799 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.service_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.system_scope = all log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.800 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.valid_interfaces = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_limit.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_reports.file_event_handler = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_reports.file_event_handler_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.801 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] oslo_reports.log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.capabilities = [12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_linux_bridge_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.capabilities = [12, 1] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.802 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] vif_plug_ovs_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.flat_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.forward_bridge_interface = ['all'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.803 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.iptables_bottom_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.iptables_drop_action = DROP log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.iptables_top_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.use_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_linux_bridge.vlan_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.isolate_vif = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.804 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.ovs_vsctl_timeout = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.ovsdb_interface = native log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_vif_ovs.per_port_bridge = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_brick.lock_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_brick.wait_mpath_device_attempts = 4 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] os_brick.wait_mpath_device_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.capabilities = [21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.805 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.logger_name = os_brick.privileged log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] privsep_osbrick.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.806 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.807 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.807 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.807 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] nova_sys_admin.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.807 231470 DEBUG oslo_service.service [None req-9259a254-2887-4735-8881-9a7dbbd18a7f - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.808 231470 INFO nova.service [-] Starting compute node (version 27.5.2-0.20250829104910.6f8decf.el9)#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.836 231470 INFO nova.virt.node [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.837 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Starting native event thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:492#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.837 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Starting green dispatch thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:498#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.838 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Starting connection event dispatch thread initialize /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:620#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.838 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Connecting to libvirt: qemu:///system _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:503#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.846 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Registering for lifecycle events _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:509#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.848 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Registering for connection events: _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:530#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.849 231470 INFO nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Connection event '1' reason 'None'#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.868 231470 INFO nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Libvirt host capabilities Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 8a8a8082-6126-4917-bb42-c4150a0ac6a5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: x86_64 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v4 Nov 27 04:34:22 localhost nova_compute[231466]: AMD Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: tcp Nov 27 04:34:22 localhost nova_compute[231466]: rdma Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 16116612 Nov 27 04:34:22 localhost nova_compute[231466]: 4029153 Nov 27 04:34:22 localhost nova_compute[231466]: 0 Nov 27 04:34:22 localhost nova_compute[231466]: 0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: selinux Nov 27 04:34:22 localhost nova_compute[231466]: 0 Nov 27 04:34:22 localhost nova_compute[231466]: system_u:system_r:svirt_t:s0 Nov 27 04:34:22 localhost nova_compute[231466]: system_u:system_r:svirt_tcg_t:s0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: dac Nov 27 04:34:22 localhost nova_compute[231466]: 0 Nov 27 04:34:22 localhost nova_compute[231466]: +107:+107 Nov 27 04:34:22 localhost nova_compute[231466]: +107:+107 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: hvm Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 32 Nov 27 04:34:22 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:22 localhost nova_compute[231466]: pc-i440fx-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.8.0 Nov 27 04:34:22 localhost nova_compute[231466]: q35 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.4.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.5.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.3.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.4.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.2.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.2.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.0.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.0.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.1.0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: hvm Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 64 Nov 27 04:34:22 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:22 localhost nova_compute[231466]: pc-i440fx-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.8.0 Nov 27 04:34:22 localhost nova_compute[231466]: q35 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.4.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.5.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.3.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.4.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.2.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.2.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.0.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.0.0 Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel8.1.0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: #033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.873 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Getting domain capabilities for i686 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.885 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=pc: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:22 localhost nova_compute[231466]: kvm Nov 27 04:34:22 localhost nova_compute[231466]: pc-i440fx-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: i686 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: rom Nov 27 04:34:22 localhost nova_compute[231466]: pflash Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: yes Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:22 localhost nova_compute[231466]: AMD Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 486 Nov 27 04:34:22 localhost nova_compute[231466]: 486-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Conroe Nov 27 04:34:22 localhost nova_compute[231466]: Conroe-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Genoa Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Genoa-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-IBPB Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v4 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v1 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v2 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v6 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v7 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: KnightsMill Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: KnightsMill-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G1-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G2 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G2-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G3 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G3-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G4-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G5-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Penryn Nov 27 04:34:22 localhost nova_compute[231466]: Penryn-v1 Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SierraForest Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SierraForest-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Westmere Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-v2 Nov 27 04:34:22 localhost nova_compute[231466]: athlon Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: athlon-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: core2duo Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: core2duo-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: coreduo Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: coreduo-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: kvm32 Nov 27 04:34:22 localhost nova_compute[231466]: kvm32-v1 Nov 27 04:34:22 localhost nova_compute[231466]: kvm64 Nov 27 04:34:22 localhost nova_compute[231466]: kvm64-v1 Nov 27 04:34:22 localhost nova_compute[231466]: n270 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: n270-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: pentium Nov 27 04:34:22 localhost nova_compute[231466]: pentium-v1 Nov 27 04:34:22 localhost nova_compute[231466]: pentium2 Nov 27 04:34:22 localhost nova_compute[231466]: pentium2-v1 Nov 27 04:34:22 localhost nova_compute[231466]: pentium3 Nov 27 04:34:22 localhost nova_compute[231466]: pentium3-v1 Nov 27 04:34:22 localhost nova_compute[231466]: phenom Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: phenom-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: qemu32 Nov 27 04:34:22 localhost nova_compute[231466]: qemu32-v1 Nov 27 04:34:22 localhost nova_compute[231466]: qemu64 Nov 27 04:34:22 localhost nova_compute[231466]: qemu64-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: file Nov 27 04:34:22 localhost nova_compute[231466]: anonymous Nov 27 04:34:22 localhost nova_compute[231466]: memfd Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: disk Nov 27 04:34:22 localhost nova_compute[231466]: cdrom Nov 27 04:34:22 localhost nova_compute[231466]: floppy Nov 27 04:34:22 localhost nova_compute[231466]: lun Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: ide Nov 27 04:34:22 localhost nova_compute[231466]: fdc Nov 27 04:34:22 localhost nova_compute[231466]: scsi Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: sata Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:22 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: vnc Nov 27 04:34:22 localhost nova_compute[231466]: egl-headless Nov 27 04:34:22 localhost nova_compute[231466]: dbus Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: subsystem Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: default Nov 27 04:34:22 localhost nova_compute[231466]: mandatory Nov 27 04:34:22 localhost nova_compute[231466]: requisite Nov 27 04:34:22 localhost nova_compute[231466]: optional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: pci Nov 27 04:34:22 localhost nova_compute[231466]: scsi Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:22 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: random Nov 27 04:34:22 localhost nova_compute[231466]: egd Nov 27 04:34:22 localhost nova_compute[231466]: builtin Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: path Nov 27 04:34:22 localhost nova_compute[231466]: handle Nov 27 04:34:22 localhost nova_compute[231466]: virtiofs Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: tpm-tis Nov 27 04:34:22 localhost nova_compute[231466]: tpm-crb Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: emulator Nov 27 04:34:22 localhost nova_compute[231466]: external Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 2.0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: pty Nov 27 04:34:22 localhost nova_compute[231466]: unix Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: qemu Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: builtin Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: default Nov 27 04:34:22 localhost nova_compute[231466]: passt Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: isa Nov 27 04:34:22 localhost nova_compute[231466]: hyperv Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: null Nov 27 04:34:22 localhost nova_compute[231466]: vc Nov 27 04:34:22 localhost nova_compute[231466]: pty Nov 27 04:34:22 localhost nova_compute[231466]: dev Nov 27 04:34:22 localhost nova_compute[231466]: file Nov 27 04:34:22 localhost nova_compute[231466]: pipe Nov 27 04:34:22 localhost nova_compute[231466]: stdio Nov 27 04:34:22 localhost nova_compute[231466]: udp Nov 27 04:34:22 localhost nova_compute[231466]: tcp Nov 27 04:34:22 localhost nova_compute[231466]: unix Nov 27 04:34:22 localhost nova_compute[231466]: qemu-vdagent Nov 27 04:34:22 localhost nova_compute[231466]: dbus Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: relaxed Nov 27 04:34:22 localhost nova_compute[231466]: vapic Nov 27 04:34:22 localhost nova_compute[231466]: spinlocks Nov 27 04:34:22 localhost nova_compute[231466]: vpindex Nov 27 04:34:22 localhost nova_compute[231466]: runtime Nov 27 04:34:22 localhost nova_compute[231466]: synic Nov 27 04:34:22 localhost nova_compute[231466]: stimer Nov 27 04:34:22 localhost nova_compute[231466]: reset Nov 27 04:34:22 localhost nova_compute[231466]: vendor_id Nov 27 04:34:22 localhost nova_compute[231466]: frequencies Nov 27 04:34:22 localhost nova_compute[231466]: reenlightenment Nov 27 04:34:22 localhost nova_compute[231466]: tlbflush Nov 27 04:34:22 localhost nova_compute[231466]: ipi Nov 27 04:34:22 localhost nova_compute[231466]: avic Nov 27 04:34:22 localhost nova_compute[231466]: emsr_bitmap Nov 27 04:34:22 localhost nova_compute[231466]: xmm_input Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 4095 Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Linux KVM Hv Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: tdx Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.888 231470 DEBUG nova.virt.libvirt.volume.mount [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Initialising _HostMountState generation 0 host_up /usr/lib/python3.9/site-packages/nova/virt/libvirt/volume/mount.py:130#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.895 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=q35: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:22 localhost nova_compute[231466]: kvm Nov 27 04:34:22 localhost nova_compute[231466]: pc-q35-rhel9.8.0 Nov 27 04:34:22 localhost nova_compute[231466]: i686 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: rom Nov 27 04:34:22 localhost nova_compute[231466]: pflash Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: yes Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:22 localhost nova_compute[231466]: AMD Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 486 Nov 27 04:34:22 localhost nova_compute[231466]: 486-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Broadwell-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cascadelake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Conroe Nov 27 04:34:22 localhost nova_compute[231466]: Conroe-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Cooperlake-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Denverton-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Dhyana-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Genoa Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Genoa-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-IBPB Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Milan-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-Rome-v4 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v1 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v2 Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: EPYC-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: GraniteRapids-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Haswell-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-noTSX Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v6 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Icelake-Server-v7 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: IvyBridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: KnightsMill Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: KnightsMill-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nehalem-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G1-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G2 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G2-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G3 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G3-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G4-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Opteron_G5-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Penryn Nov 27 04:34:22 localhost nova_compute[231466]: Penryn-v1 Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: SandyBridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SapphireRapids-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SierraForest Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: SierraForest-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Client-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-noTSX-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Skylake-Server-v5 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v2 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v3 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Snowridge-v4 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Westmere Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-IBRS Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Westmere-v2 Nov 27 04:34:22 localhost nova_compute[231466]: athlon Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: athlon-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: core2duo Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: core2duo-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: coreduo Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: coreduo-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: kvm32 Nov 27 04:34:22 localhost nova_compute[231466]: kvm32-v1 Nov 27 04:34:22 localhost nova_compute[231466]: kvm64 Nov 27 04:34:22 localhost nova_compute[231466]: kvm64-v1 Nov 27 04:34:22 localhost nova_compute[231466]: n270 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: n270-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: pentium Nov 27 04:34:22 localhost nova_compute[231466]: pentium-v1 Nov 27 04:34:22 localhost nova_compute[231466]: pentium2 Nov 27 04:34:22 localhost nova_compute[231466]: pentium2-v1 Nov 27 04:34:22 localhost nova_compute[231466]: pentium3 Nov 27 04:34:22 localhost nova_compute[231466]: pentium3-v1 Nov 27 04:34:22 localhost nova_compute[231466]: phenom Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: phenom-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: qemu32 Nov 27 04:34:22 localhost nova_compute[231466]: qemu32-v1 Nov 27 04:34:22 localhost nova_compute[231466]: qemu64 Nov 27 04:34:22 localhost nova_compute[231466]: qemu64-v1 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: file Nov 27 04:34:22 localhost nova_compute[231466]: anonymous Nov 27 04:34:22 localhost nova_compute[231466]: memfd Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: disk Nov 27 04:34:22 localhost nova_compute[231466]: cdrom Nov 27 04:34:22 localhost nova_compute[231466]: floppy Nov 27 04:34:22 localhost nova_compute[231466]: lun Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: fdc Nov 27 04:34:22 localhost nova_compute[231466]: scsi Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: sata Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:22 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: vnc Nov 27 04:34:22 localhost nova_compute[231466]: egl-headless Nov 27 04:34:22 localhost nova_compute[231466]: dbus Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: subsystem Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: default Nov 27 04:34:22 localhost nova_compute[231466]: mandatory Nov 27 04:34:22 localhost nova_compute[231466]: requisite Nov 27 04:34:22 localhost nova_compute[231466]: optional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: pci Nov 27 04:34:22 localhost nova_compute[231466]: scsi Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: virtio Nov 27 04:34:22 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:22 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: random Nov 27 04:34:22 localhost nova_compute[231466]: egd Nov 27 04:34:22 localhost nova_compute[231466]: builtin Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: path Nov 27 04:34:22 localhost nova_compute[231466]: handle Nov 27 04:34:22 localhost nova_compute[231466]: virtiofs Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: tpm-tis Nov 27 04:34:22 localhost nova_compute[231466]: tpm-crb Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: emulator Nov 27 04:34:22 localhost nova_compute[231466]: external Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 2.0 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: usb Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: pty Nov 27 04:34:22 localhost nova_compute[231466]: unix Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: qemu Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: builtin Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: default Nov 27 04:34:22 localhost nova_compute[231466]: passt Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: isa Nov 27 04:34:22 localhost nova_compute[231466]: hyperv Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: null Nov 27 04:34:22 localhost nova_compute[231466]: vc Nov 27 04:34:22 localhost nova_compute[231466]: pty Nov 27 04:34:22 localhost nova_compute[231466]: dev Nov 27 04:34:22 localhost nova_compute[231466]: file Nov 27 04:34:22 localhost nova_compute[231466]: pipe Nov 27 04:34:22 localhost nova_compute[231466]: stdio Nov 27 04:34:22 localhost nova_compute[231466]: udp Nov 27 04:34:22 localhost nova_compute[231466]: tcp Nov 27 04:34:22 localhost nova_compute[231466]: unix Nov 27 04:34:22 localhost nova_compute[231466]: qemu-vdagent Nov 27 04:34:22 localhost nova_compute[231466]: dbus Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: relaxed Nov 27 04:34:22 localhost nova_compute[231466]: vapic Nov 27 04:34:22 localhost nova_compute[231466]: spinlocks Nov 27 04:34:22 localhost nova_compute[231466]: vpindex Nov 27 04:34:22 localhost nova_compute[231466]: runtime Nov 27 04:34:22 localhost nova_compute[231466]: synic Nov 27 04:34:22 localhost nova_compute[231466]: stimer Nov 27 04:34:22 localhost nova_compute[231466]: reset Nov 27 04:34:22 localhost nova_compute[231466]: vendor_id Nov 27 04:34:22 localhost nova_compute[231466]: frequencies Nov 27 04:34:22 localhost nova_compute[231466]: reenlightenment Nov 27 04:34:22 localhost nova_compute[231466]: tlbflush Nov 27 04:34:22 localhost nova_compute[231466]: ipi Nov 27 04:34:22 localhost nova_compute[231466]: avic Nov 27 04:34:22 localhost nova_compute[231466]: emsr_bitmap Nov 27 04:34:22 localhost nova_compute[231466]: xmm_input Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: 4095 Nov 27 04:34:22 localhost nova_compute[231466]: on Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: off Nov 27 04:34:22 localhost nova_compute[231466]: Linux KVM Hv Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: tdx Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.915 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Getting domain capabilities for x86_64 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:34:22 localhost nova_compute[231466]: 2025-11-27 09:34:22.920 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=pc: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:22 localhost nova_compute[231466]: kvm Nov 27 04:34:22 localhost nova_compute[231466]: pc-i440fx-rhel7.6.0 Nov 27 04:34:22 localhost nova_compute[231466]: x86_64 Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: rom Nov 27 04:34:22 localhost nova_compute[231466]: pflash Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: yes Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: no Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:22 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:23 localhost nova_compute[231466]: AMD Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 486 Nov 27 04:34:23 localhost nova_compute[231466]: 486-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Conroe Nov 27 04:34:23 localhost nova_compute[231466]: Conroe-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Genoa Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Genoa-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-IBPB Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v4 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v1 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v2 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v6 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v7 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: KnightsMill Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: KnightsMill-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G1-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G2 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G2-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G3 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G3-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G4-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G5-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Penryn Nov 27 04:34:23 localhost nova_compute[231466]: Penryn-v1 Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SierraForest Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SierraForest-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Westmere Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-v2 Nov 27 04:34:23 localhost nova_compute[231466]: athlon Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: athlon-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: core2duo Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: core2duo-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: coreduo Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: coreduo-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: kvm32 Nov 27 04:34:23 localhost nova_compute[231466]: kvm32-v1 Nov 27 04:34:23 localhost nova_compute[231466]: kvm64 Nov 27 04:34:23 localhost nova_compute[231466]: kvm64-v1 Nov 27 04:34:23 localhost nova_compute[231466]: n270 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: n270-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: pentium Nov 27 04:34:23 localhost nova_compute[231466]: pentium-v1 Nov 27 04:34:23 localhost nova_compute[231466]: pentium2 Nov 27 04:34:23 localhost nova_compute[231466]: pentium2-v1 Nov 27 04:34:23 localhost nova_compute[231466]: pentium3 Nov 27 04:34:23 localhost nova_compute[231466]: pentium3-v1 Nov 27 04:34:23 localhost nova_compute[231466]: phenom Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: phenom-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: qemu32 Nov 27 04:34:23 localhost nova_compute[231466]: qemu32-v1 Nov 27 04:34:23 localhost nova_compute[231466]: qemu64 Nov 27 04:34:23 localhost nova_compute[231466]: qemu64-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: file Nov 27 04:34:23 localhost nova_compute[231466]: anonymous Nov 27 04:34:23 localhost nova_compute[231466]: memfd Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: disk Nov 27 04:34:23 localhost nova_compute[231466]: cdrom Nov 27 04:34:23 localhost nova_compute[231466]: floppy Nov 27 04:34:23 localhost nova_compute[231466]: lun Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: ide Nov 27 04:34:23 localhost nova_compute[231466]: fdc Nov 27 04:34:23 localhost nova_compute[231466]: scsi Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: sata Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:23 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: vnc Nov 27 04:34:23 localhost nova_compute[231466]: egl-headless Nov 27 04:34:23 localhost nova_compute[231466]: dbus Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: subsystem Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: default Nov 27 04:34:23 localhost nova_compute[231466]: mandatory Nov 27 04:34:23 localhost nova_compute[231466]: requisite Nov 27 04:34:23 localhost nova_compute[231466]: optional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: pci Nov 27 04:34:23 localhost nova_compute[231466]: scsi Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:23 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: random Nov 27 04:34:23 localhost nova_compute[231466]: egd Nov 27 04:34:23 localhost nova_compute[231466]: builtin Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: path Nov 27 04:34:23 localhost nova_compute[231466]: handle Nov 27 04:34:23 localhost nova_compute[231466]: virtiofs Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: tpm-tis Nov 27 04:34:23 localhost nova_compute[231466]: tpm-crb Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: emulator Nov 27 04:34:23 localhost nova_compute[231466]: external Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 2.0 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: pty Nov 27 04:34:23 localhost nova_compute[231466]: unix Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: qemu Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: builtin Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: default Nov 27 04:34:23 localhost nova_compute[231466]: passt Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: isa Nov 27 04:34:23 localhost nova_compute[231466]: hyperv Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: null Nov 27 04:34:23 localhost nova_compute[231466]: vc Nov 27 04:34:23 localhost nova_compute[231466]: pty Nov 27 04:34:23 localhost nova_compute[231466]: dev Nov 27 04:34:23 localhost nova_compute[231466]: file Nov 27 04:34:23 localhost nova_compute[231466]: pipe Nov 27 04:34:23 localhost nova_compute[231466]: stdio Nov 27 04:34:23 localhost nova_compute[231466]: udp Nov 27 04:34:23 localhost nova_compute[231466]: tcp Nov 27 04:34:23 localhost nova_compute[231466]: unix Nov 27 04:34:23 localhost nova_compute[231466]: qemu-vdagent Nov 27 04:34:23 localhost nova_compute[231466]: dbus Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: relaxed Nov 27 04:34:23 localhost nova_compute[231466]: vapic Nov 27 04:34:23 localhost nova_compute[231466]: spinlocks Nov 27 04:34:23 localhost nova_compute[231466]: vpindex Nov 27 04:34:23 localhost nova_compute[231466]: runtime Nov 27 04:34:23 localhost nova_compute[231466]: synic Nov 27 04:34:23 localhost nova_compute[231466]: stimer Nov 27 04:34:23 localhost nova_compute[231466]: reset Nov 27 04:34:23 localhost nova_compute[231466]: vendor_id Nov 27 04:34:23 localhost nova_compute[231466]: frequencies Nov 27 04:34:23 localhost nova_compute[231466]: reenlightenment Nov 27 04:34:23 localhost nova_compute[231466]: tlbflush Nov 27 04:34:23 localhost nova_compute[231466]: ipi Nov 27 04:34:23 localhost nova_compute[231466]: avic Nov 27 04:34:23 localhost nova_compute[231466]: emsr_bitmap Nov 27 04:34:23 localhost nova_compute[231466]: xmm_input Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 4095 Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Linux KVM Hv Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: tdx Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:22.975 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=q35: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: /usr/libexec/qemu-kvm Nov 27 04:34:23 localhost nova_compute[231466]: kvm Nov 27 04:34:23 localhost nova_compute[231466]: pc-q35-rhel9.8.0 Nov 27 04:34:23 localhost nova_compute[231466]: x86_64 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: efi Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: /usr/share/edk2/ovmf/OVMF_CODE.secboot.fd Nov 27 04:34:23 localhost nova_compute[231466]: /usr/share/edk2/ovmf/OVMF_CODE.fd Nov 27 04:34:23 localhost nova_compute[231466]: /usr/share/edk2/ovmf/OVMF.amdsev.fd Nov 27 04:34:23 localhost nova_compute[231466]: /usr/share/edk2/ovmf/OVMF.inteltdx.secboot.fd Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: rom Nov 27 04:34:23 localhost nova_compute[231466]: pflash Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: yes Nov 27 04:34:23 localhost nova_compute[231466]: no Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: yes Nov 27 04:34:23 localhost nova_compute[231466]: no Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:23 localhost nova_compute[231466]: AMD Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 486 Nov 27 04:34:23 localhost nova_compute[231466]: 486-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Broadwell-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cascadelake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Conroe Nov 27 04:34:23 localhost nova_compute[231466]: Conroe-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Cooperlake-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Denverton-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Dhyana-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Genoa Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Genoa-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-IBPB Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Milan-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-Rome-v4 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v1 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v2 Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: EPYC-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: GraniteRapids-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Haswell-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-noTSX Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v6 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Icelake-Server-v7 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: IvyBridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: KnightsMill Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: KnightsMill-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nehalem-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G1-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G2 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G2-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G3 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G3-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G4-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Opteron_G5-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Penryn Nov 27 04:34:23 localhost nova_compute[231466]: Penryn-v1 Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: SandyBridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SapphireRapids-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SierraForest Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: SierraForest-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Client-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-noTSX-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Skylake-Server-v5 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v2 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v3 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Snowridge-v4 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Westmere Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-IBRS Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Westmere-v2 Nov 27 04:34:23 localhost nova_compute[231466]: athlon Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: athlon-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: core2duo Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: core2duo-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: coreduo Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: coreduo-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: kvm32 Nov 27 04:34:23 localhost nova_compute[231466]: kvm32-v1 Nov 27 04:34:23 localhost nova_compute[231466]: kvm64 Nov 27 04:34:23 localhost nova_compute[231466]: kvm64-v1 Nov 27 04:34:23 localhost nova_compute[231466]: n270 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: n270-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: pentium Nov 27 04:34:23 localhost nova_compute[231466]: pentium-v1 Nov 27 04:34:23 localhost nova_compute[231466]: pentium2 Nov 27 04:34:23 localhost nova_compute[231466]: pentium2-v1 Nov 27 04:34:23 localhost nova_compute[231466]: pentium3 Nov 27 04:34:23 localhost nova_compute[231466]: pentium3-v1 Nov 27 04:34:23 localhost nova_compute[231466]: phenom Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: phenom-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: qemu32 Nov 27 04:34:23 localhost nova_compute[231466]: qemu32-v1 Nov 27 04:34:23 localhost nova_compute[231466]: qemu64 Nov 27 04:34:23 localhost nova_compute[231466]: qemu64-v1 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: file Nov 27 04:34:23 localhost nova_compute[231466]: anonymous Nov 27 04:34:23 localhost nova_compute[231466]: memfd Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: disk Nov 27 04:34:23 localhost nova_compute[231466]: cdrom Nov 27 04:34:23 localhost nova_compute[231466]: floppy Nov 27 04:34:23 localhost nova_compute[231466]: lun Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: fdc Nov 27 04:34:23 localhost nova_compute[231466]: scsi Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: sata Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:23 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: vnc Nov 27 04:34:23 localhost nova_compute[231466]: egl-headless Nov 27 04:34:23 localhost nova_compute[231466]: dbus Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: subsystem Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: default Nov 27 04:34:23 localhost nova_compute[231466]: mandatory Nov 27 04:34:23 localhost nova_compute[231466]: requisite Nov 27 04:34:23 localhost nova_compute[231466]: optional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: pci Nov 27 04:34:23 localhost nova_compute[231466]: scsi Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: virtio Nov 27 04:34:23 localhost nova_compute[231466]: virtio-transitional Nov 27 04:34:23 localhost nova_compute[231466]: virtio-non-transitional Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: random Nov 27 04:34:23 localhost nova_compute[231466]: egd Nov 27 04:34:23 localhost nova_compute[231466]: builtin Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: path Nov 27 04:34:23 localhost nova_compute[231466]: handle Nov 27 04:34:23 localhost nova_compute[231466]: virtiofs Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: tpm-tis Nov 27 04:34:23 localhost nova_compute[231466]: tpm-crb Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: emulator Nov 27 04:34:23 localhost nova_compute[231466]: external Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 2.0 Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: usb Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: pty Nov 27 04:34:23 localhost nova_compute[231466]: unix Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: qemu Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: builtin Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: default Nov 27 04:34:23 localhost nova_compute[231466]: passt Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: isa Nov 27 04:34:23 localhost nova_compute[231466]: hyperv Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: null Nov 27 04:34:23 localhost nova_compute[231466]: vc Nov 27 04:34:23 localhost nova_compute[231466]: pty Nov 27 04:34:23 localhost nova_compute[231466]: dev Nov 27 04:34:23 localhost nova_compute[231466]: file Nov 27 04:34:23 localhost nova_compute[231466]: pipe Nov 27 04:34:23 localhost nova_compute[231466]: stdio Nov 27 04:34:23 localhost nova_compute[231466]: udp Nov 27 04:34:23 localhost nova_compute[231466]: tcp Nov 27 04:34:23 localhost nova_compute[231466]: unix Nov 27 04:34:23 localhost nova_compute[231466]: qemu-vdagent Nov 27 04:34:23 localhost nova_compute[231466]: dbus Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: relaxed Nov 27 04:34:23 localhost nova_compute[231466]: vapic Nov 27 04:34:23 localhost nova_compute[231466]: spinlocks Nov 27 04:34:23 localhost nova_compute[231466]: vpindex Nov 27 04:34:23 localhost nova_compute[231466]: runtime Nov 27 04:34:23 localhost nova_compute[231466]: synic Nov 27 04:34:23 localhost nova_compute[231466]: stimer Nov 27 04:34:23 localhost nova_compute[231466]: reset Nov 27 04:34:23 localhost nova_compute[231466]: vendor_id Nov 27 04:34:23 localhost nova_compute[231466]: frequencies Nov 27 04:34:23 localhost nova_compute[231466]: reenlightenment Nov 27 04:34:23 localhost nova_compute[231466]: tlbflush Nov 27 04:34:23 localhost nova_compute[231466]: ipi Nov 27 04:34:23 localhost nova_compute[231466]: avic Nov 27 04:34:23 localhost nova_compute[231466]: emsr_bitmap Nov 27 04:34:23 localhost nova_compute[231466]: xmm_input Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: 4095 Nov 27 04:34:23 localhost nova_compute[231466]: on Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: off Nov 27 04:34:23 localhost nova_compute[231466]: Linux KVM Hv Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: tdx Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: Nov 27 04:34:23 localhost nova_compute[231466]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.032 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.032 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.032 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.033 231470 INFO nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Secure Boot support detected#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.034 231470 INFO nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.035 231470 INFO nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.044 231470 DEBUG nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Enabling emulated TPM support _check_vtpm_support /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:1097#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.079 231470 INFO nova.virt.node [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.100 231470 DEBUG nova.compute.manager [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Verified node 5764ad4c-cf89-4d5f-a185-92be71f7b67f matches my host np0005537446.localdomain _check_for_host_rename /usr/lib/python3.9/site-packages/nova/compute/manager.py:1568#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.139 231470 DEBUG nova.compute.manager [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.142 231470 DEBUG nova.virt.libvirt.vif [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=,hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=,node='np0005537446.localdomain',numa_topology=None,old_flavor=,os_type=None,pci_devices=,pci_requests=,power_state=1,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata=,tags=,task_state=None,terminated_at=None,trusted_certs=,updated_at=2025-11-27T08:33:11Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"dhcp_server": "192.168.0.1"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system"}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.142 231470 DEBUG nova.network.os_vif_util [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"dhcp_server": "192.168.0.1"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system"}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.143 231470 DEBUG nova.network.os_vif_util [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Converted object VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.143 231470 DEBUG os_vif [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Plugging vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.166 231470 DEBUG ovsdbapp.backend.ovs_idl [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Created schema index Interface.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.166 231470 DEBUG ovsdbapp.backend.ovs_idl [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Created schema index Port.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.167 231470 DEBUG ovsdbapp.backend.ovs_idl [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Created schema index Bridge.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.167 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] tcp:127.0.0.1:6640: entering CONNECTING _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.167 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] [POLLOUT] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.167 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.168 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.181 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.181 231470 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.181 231470 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.182 231470 INFO oslo.privsep.daemon [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Running privsep helper: ['sudo', 'nova-rootwrap', '/etc/nova/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-compute.conf', '--config-dir', '/etc/nova/nova.conf.d', '--privsep_context', 'vif_plug_ovs.privsep.vif_plug', '--privsep_sock_path', '/tmp/tmpbcv8z1i_/privsep.sock']#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.488 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62129 DF PROTO=TCP SPT=51296 DPT=9101 SEQ=2673599764 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294509A0000000001030307) Nov 27 04:34:23 localhost python3.9[231809]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner.service follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.821 231470 INFO oslo.privsep.daemon [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.725 231826 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.730 231826 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.733 231826 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_DAC_OVERRIDE|CAP_NET_ADMIN/CAP_DAC_OVERRIDE|CAP_NET_ADMIN/none#033[00m Nov 27 04:34:23 localhost nova_compute[231466]: 2025-11-27 09:34:23.733 231826 INFO oslo.privsep.daemon [-] privsep daemon running as pid 231826#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.075 231470 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.076 231470 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tapa2718872-3b, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.076 231470 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tapa2718872-3b, col_values=(('external_ids', {'iface-id': 'a2718872-3b82-44ca-8d67-a0e516b2708b', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:86:b9:b3', 'vm-uuid': 'a02f7f2f-d2cf-4612-b18a-8be435257201'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.077 231470 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.077 231470 INFO os_vif [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Successfully plugged vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b')#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.078 231470 DEBUG nova.compute.manager [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.081 231470 DEBUG nova.compute.manager [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Current state is 1, state in DB is 1. _init_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:1304#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.081 231470 INFO nova.compute.manager [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host#033[00m Nov 27 04:34:24 localhost python3.9[231954]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner.service.requires follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.615 231470 INFO nova.service [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating service version for nova-compute on np0005537446.localdomain from 57 to 66#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.805 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.806 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.806 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.806 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:34:24 localhost nova_compute[231466]: 2025-11-27 09:34:24.806 231470 DEBUG oslo_concurrency.processutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.260 231470 DEBUG oslo_concurrency.processutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.454s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.526 231470 DEBUG nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.526 231470 DEBUG nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:34:25 localhost systemd[1]: Started libvirt nodedev daemon. Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.914 231470 WARNING nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.915 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=12925MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.915 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:34:25 localhost nova_compute[231466]: 2025-11-27 09:34:25.916 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:34:26 localhost python3.9[232127]: ansible-containers.podman.podman_container Invoked with name=nova_nvme_cleaner state=absent executable=podman detach=True debug=False force_restart=False force_delete=True generate_systemd={} image_strict=False recreate=False image=None annotation=None arch=None attach=None authfile=None blkio_weight=None blkio_weight_device=None cap_add=None cap_drop=None cgroup_conf=None cgroup_parent=None cgroupns=None cgroups=None chrootdirs=None cidfile=None cmd_args=None conmon_pidfile=None command=None cpu_period=None cpu_quota=None cpu_rt_period=None cpu_rt_runtime=None cpu_shares=None cpus=None cpuset_cpus=None cpuset_mems=None decryption_key=None delete_depend=None delete_time=None delete_volumes=None detach_keys=None device=None device_cgroup_rule=None device_read_bps=None device_read_iops=None device_write_bps=None device_write_iops=None dns=None dns_option=None dns_search=None entrypoint=None env=None env_file=None env_host=None env_merge=None etc_hosts=None expose=None gidmap=None gpus=None group_add=None group_entry=None healthcheck=None healthcheck_interval=None healthcheck_retries=None healthcheck_start_period=None health_startup_cmd=None health_startup_interval=None health_startup_retries=None health_startup_success=None health_startup_timeout=None healthcheck_timeout=None healthcheck_failure_action=None hooks_dir=None hostname=None hostuser=None http_proxy=None image_volume=None init=None init_ctr=None init_path=None interactive=None ip=None ip6=None ipc=None kernel_memory=None label=None label_file=None log_driver=None log_level=None log_opt=None mac_address=None memory=None memory_reservation=None memory_swap=None memory_swappiness=None mount=None network=None network_aliases=None no_healthcheck=None no_hosts=None oom_kill_disable=None oom_score_adj=None os=None passwd=None passwd_entry=None personality=None pid=None pid_file=None pids_limit=None platform=None pod=None pod_id_file=None preserve_fd=None preserve_fds=None privileged=None publish=None publish_all=None pull=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None rdt_class=None read_only=None read_only_tmpfs=None requires=None restart_policy=None restart_time=None retry=None retry_delay=None rm=None rmi=None rootfs=None seccomp_policy=None secrets=NOT_LOGGING_PARAMETER sdnotify=None security_opt=None shm_size=None shm_size_systemd=None sig_proxy=None stop_signal=None stop_timeout=None stop_time=None subgidname=None subuidname=None sysctl=None systemd=None timeout=None timezone=None tls_verify=None tmpfs=None tty=None uidmap=None ulimit=None umask=None unsetenv=None unsetenv_all=None user=None userns=None uts=None variant=None volume=None volumes_from=None workdir=None Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.207 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:34:26 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 115.0 (383 of 333 items), suggesting rotation. Nov 27 04:34:26 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.207 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:34:26 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.207 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.221 231470 DEBUG nova.scheduler.client.report [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:34:26 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.316 231470 DEBUG nova.scheduler.client.report [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.317 231470 DEBUG nova.compute.provider_tree [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.337 231470 DEBUG nova.scheduler.client.report [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.378 231470 DEBUG nova.scheduler.client.report [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_F16C,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,HW_CPU_X86_SSE4A,HW_CPU_X86_MMX,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_VIRTIO,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_CIRRUS,COMPUTE_NET_VIF_MODEL_E1000,HW_CPU_X86_ABM,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_STORAGE_BUS_FDC,COMPUTE_STORAGE_BUS_SCSI,COMPUTE_NET_VIF_MODEL_LAN9118,COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_VIF_MODEL_RTL8139,HW_CPU_X86_BMI,COMPUTE_DEVICE_TAGGING,HW_CPU_X86_BMI2,COMPUTE_STORAGE_BUS_USB,COMPUTE_TRUSTED_CERTS,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SSSE3,HW_CPU_X86_FMA3,COMPUTE_RESCUE_BFV,HW_CPU_X86_AESNI,HW_CPU_X86_SVM,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_STORAGE_BUS_SATA,COMPUTE_NET_VIF_MODEL_VIRTIO,HW_CPU_X86_AMD_SVM,COMPUTE_NET_VIF_MODEL_E1000E,COMPUTE_NET_VIF_MODEL_PCNET,COMPUTE_GRAPHICS_MODEL_VGA,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_NODE,COMPUTE_VOLUME_EXTEND,HW_CPU_X86_SSE42,COMPUTE_ACCELERATORS,COMPUTE_STORAGE_BUS_IDE,COMPUTE_GRAPHICS_MODEL_NONE,HW_CPU_X86_CLMUL,HW_CPU_X86_SSE,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_AVX2,HW_CPU_X86_SSE41,HW_CPU_X86_SSE2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.425 231470 DEBUG oslo_concurrency.processutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20748 DF PROTO=TCP SPT=32890 DPT=9105 SEQ=2626203627 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12945CDB0000000001030307) Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.903 231470 DEBUG oslo_concurrency.processutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.477s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.909 231470 DEBUG nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] /sys/module/kvm_amd/parameters/sev contains [N Nov 27 04:34:26 localhost nova_compute[231466]: ] _kernel_supports_amd_sev /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1803#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.910 231470 INFO nova.virt.libvirt.host [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] kernel doesn't support AMD SEV#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.911 231470 DEBUG nova.compute.provider_tree [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'MEMORY_MB': {'total': 15738, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 512}, 'VCPU': {'total': 8, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0, 'reserved': 0}, 'DISK_GB': {'total': 41, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 1}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.912 231470 DEBUG nova.virt.libvirt.driver [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] CPU mode 'host-model' models '' was chosen, with extra flags: '' _get_guest_cpu_model_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:5396#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.992 231470 DEBUG nova.scheduler.client.report [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updated inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with generation 3 in Placement from set_inventory_for_provider using data: {'MEMORY_MB': {'total': 15738, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 512}, 'VCPU': {'total': 8, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0, 'reserved': 0}, 'DISK_GB': {'total': 41, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 1}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:957#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.993 231470 DEBUG nova.compute.provider_tree [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f generation from 3 to 4 during operation: update_inventory _update_generation /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:164#033[00m Nov 27 04:34:26 localhost nova_compute[231466]: 2025-11-27 09:34:26.993 231470 DEBUG nova.compute.provider_tree [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.054 231470 DEBUG nova.compute.provider_tree [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Updating resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f generation from 4 to 5 during operation: update_traits _update_generation /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:164#033[00m Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.104 231470 DEBUG nova.compute.resource_tracker [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.104 231470 DEBUG oslo_concurrency.lockutils [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.188s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.104 231470 DEBUG nova.service [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Creating RPC server for service compute start /usr/lib/python3.9/site-packages/nova/service.py:182#033[00m Nov 27 04:34:27 localhost python3.9[232280]: ansible-ansible.builtin.systemd Invoked with name=edpm_nova_compute.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:34:27 localhost systemd[1]: Stopping nova_compute container... Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.229 231470 DEBUG nova.service [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] Join ServiceGroup membership for this service compute start /usr/lib/python3.9/site-packages/nova/service.py:199#033[00m Nov 27 04:34:27 localhost nova_compute[231466]: 2025-11-27 09:34:27.230 231470 DEBUG nova.servicegroup.drivers.db [None req-ca516551-8bae-4227-90fb-2d48184bf9ef - - - - - -] DB_Driver: join new ServiceGroup member np0005537446.localdomain to the compute group, service = join /usr/lib/python3.9/site-packages/nova/servicegroup/drivers/db.py:44#033[00m Nov 27 04:34:27 localhost journal[205316]: libvirt version: 11.9.0, package: 1.el9 (builder@centos.org, 2025-11-04-09:54:50, ) Nov 27 04:34:27 localhost journal[205316]: hostname: np0005537446.localdomain Nov 27 04:34:27 localhost journal[205316]: End of file while reading data: Input/output error Nov 27 04:34:27 localhost systemd[1]: libpod-4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6.scope: Deactivated successfully. Nov 27 04:34:27 localhost systemd[1]: libpod-4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6.scope: Consumed 4.493s CPU time. Nov 27 04:34:27 localhost podman[232286]: 2025-11-27 09:34:27.272939437 +0000 UTC m=+0.104798362 container died 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, container_name=nova_compute, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 04:34:27 localhost systemd[1]: tmp-crun.ssPcia.mount: Deactivated successfully. Nov 27 04:34:28 localhost systemd[1]: var-lib-containers-storage-overlay-be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c-merged.mount: Deactivated successfully. Nov 27 04:34:28 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6-userdata-shm.mount: Deactivated successfully. Nov 27 04:34:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:34:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=1893 DF PROTO=TCP SPT=53950 DPT=9105 SEQ=1044711163 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129467DA0000000001030307) Nov 27 04:34:29 localhost systemd[1]: tmp-crun.llS0YI.mount: Deactivated successfully. Nov 27 04:34:29 localhost podman[232569]: 2025-11-27 09:34:29.980894774 +0000 UTC m=+1.074285876 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 04:34:30 localhost podman[232569]: 2025-11-27 09:34:30.290056644 +0000 UTC m=+1.383447686 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:34:32 localhost podman[232286]: 2025-11-27 09:34:32.027790428 +0000 UTC m=+4.859649313 container cleanup 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, org.label-schema.name=CentOS Stream 9 Base Image, container_name=nova_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 04:34:32 localhost podman[232286]: nova_compute Nov 27 04:34:32 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:34:32 localhost podman[232602]: error opening file `/run/crun/4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6/status`: No such file or directory Nov 27 04:34:32 localhost podman[232590]: 2025-11-27 09:34:32.131464359 +0000 UTC m=+0.067974827 container cleanup 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=nova_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, config_id=edpm, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:34:32 localhost podman[232590]: nova_compute Nov 27 04:34:32 localhost systemd[1]: edpm_nova_compute.service: Deactivated successfully. Nov 27 04:34:32 localhost systemd[1]: Stopped nova_compute container. Nov 27 04:34:32 localhost systemd[1]: Starting nova_compute container... Nov 27 04:34:32 localhost systemd[1]: Started libcrun container. Nov 27 04:34:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/nvme supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/multipath supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:32 localhost podman[232604]: 2025-11-27 09:34:32.31480542 +0000 UTC m=+0.149778127 container init 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, container_name=nova_compute, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, config_id=edpm, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:34:32 localhost podman[232604]: 2025-11-27 09:34:32.324704767 +0000 UTC m=+0.159677474 container start 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, container_name=nova_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 04:34:32 localhost podman[232604]: nova_compute Nov 27 04:34:32 localhost nova_compute[232618]: + sudo -E kolla_set_configs Nov 27 04:34:32 localhost systemd[1]: Started nova_compute container. Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Validating config file Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying service configuration files Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/01-nova.conf to /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/03-ceph-nova.conf to /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/99-nova-compute-cells-workarounds.conf to /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/02-nova-host-specific.conf to /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /etc/ceph Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Creating directory /etc/ceph Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/ceph Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.conf to /etc/ceph/ceph.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.client.openstack.keyring to /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-privatekey to /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /var/lib/nova/.ssh/config Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-config to /var/lib/nova/.ssh/config Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Deleting /usr/sbin/iscsiadm Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Copying /var/lib/kolla/config_files/run-on-host to /usr/sbin/iscsiadm Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /usr/sbin/iscsiadm Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Writing out command to execute Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:32 localhost nova_compute[232618]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:34:32 localhost nova_compute[232618]: ++ cat /run_command Nov 27 04:34:32 localhost nova_compute[232618]: + CMD=nova-compute Nov 27 04:34:32 localhost nova_compute[232618]: + ARGS= Nov 27 04:34:32 localhost nova_compute[232618]: + sudo kolla_copy_cacerts Nov 27 04:34:32 localhost nova_compute[232618]: + [[ ! -n '' ]] Nov 27 04:34:32 localhost nova_compute[232618]: + . kolla_extend_start Nov 27 04:34:32 localhost nova_compute[232618]: Running command: 'nova-compute' Nov 27 04:34:32 localhost nova_compute[232618]: + echo 'Running command: '\''nova-compute'\''' Nov 27 04:34:32 localhost nova_compute[232618]: + umask 0022 Nov 27 04:34:32 localhost nova_compute[232618]: + exec nova-compute Nov 27 04:34:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20750 DF PROTO=TCP SPT=32890 DPT=9105 SEQ=2626203627 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294749B0000000001030307) Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.127 232622 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.127 232622 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.128 232622 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.128 232622 INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.246 232622 DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.252 232622 DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 1 in 0.006s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.252 232622 DEBUG oslo_concurrency.processutils [-] 'grep -F node.session.scan /sbin/iscsiadm' failed. Not Retrying. execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:473#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.642 232622 INFO nova.virt.driver [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Loading compute driver 'libvirt.LibvirtDriver'#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.739 232622 INFO nova.compute.provider_config [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access.#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.749 232622 WARNING nova.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Current Nova version does not support computes older than Yoga but the minimum compute service level in your cell is 57 and the oldest supported service level is 61.: nova.exception.TooOldComputeService: Current Nova version does not support computes older than Yoga but the minimum compute service level in your cell is 57 and the oldest supported service level is 61.#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.749 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.749 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.750 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.750 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Full set of CONF: _wait_for_exit_or_signal /usr/lib/python3.9/site-packages/oslo_service/service.py:362#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.750 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.750 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.750 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] command line args: [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] allow_resize_to_same_host = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] arq_binding_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] backdoor_port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] backdoor_socket = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] block_device_allocate_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.751 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] block_device_allocate_retries_interval = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cert = self.pem log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute_driver = libvirt.LibvirtDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute_monitors = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] config_dir = ['/etc/nova/nova.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] config_drive_format = iso9660 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.752 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] console_host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] control_exchange = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cpu_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] default_access_ip_network_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] default_availability_zone = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.753 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] default_ephemeral_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] default_schedule_zone = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] disk_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] enable_new_services = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] enabled_apis = ['osapi_compute', 'metadata'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] enabled_ssl_apis = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.754 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] flat_injected = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] force_config_drive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] force_raw_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] heal_instance_info_cache_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] initial_cpu_allocation_ratio = 4.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.755 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] initial_disk_allocation_ratio = 0.9 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] initial_ram_allocation_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] injected_network_template = /usr/lib/python3.9/site-packages/nova/virt/interfaces.template log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_build_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_delete_interval = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_name_template = instance-%08x log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.756 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_usage_audit = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_usage_audit_period = month log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] instances_path = /var/lib/nova/instances log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] internal_service_availability_zone = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] live_migration_retry_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.757 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] log_rotation_type = size log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.758 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] long_rpc_timeout = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_concurrent_builds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_concurrent_live_migrations = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.759 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_concurrent_snapshots = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_local_block_devices = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_logfile_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] max_logfile_size_mb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] maximum_instance_delete_attempts = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metadata_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metadata_listen_port = 8775 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metadata_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.760 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] migrate_max_retries = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] mkisofs_cmd = /usr/bin/mkisofs log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] my_block_storage_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] my_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] network_allocate_retries = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] osapi_compute_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.761 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] osapi_compute_listen_port = 8774 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] osapi_compute_unique_server_name_scope = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] osapi_compute_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] password_length = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] periodic_enable = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] periodic_fuzzy_delay = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] pointer_model = usbtablet log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] preallocate_images = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.762 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] pybasedir = /usr/lib/python3.9/site-packages log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ram_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reboot_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reclaim_instance_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.763 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] record = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reimage_timeout_per_gb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] report_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rescue_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reserved_host_cpus = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reserved_host_disk_mb = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reserved_host_memory_mb = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] reserved_huge_pages = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.764 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] resize_confirm_window = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] resize_fs_using_block_device = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] resume_guests_state_on_host_boot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rootwrap_config = /etc/nova/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rpc_response_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] run_external_periodic_tasks = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] running_deleted_instance_action = reap log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.765 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] running_deleted_instance_poll_interval = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] running_deleted_instance_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler_instance_sync_interval = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_down_time = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] servicegroup_driver = db log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] shelved_offload_time = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] shelved_poll_interval = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.766 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] source_is_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ssl_only = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] state_path = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] sync_power_state_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] sync_power_state_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] tempdir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] timeout_nbd = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.767 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] update_resources_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_cow_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_rootwrap_daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.768 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vcpu_pin_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plugging_is_fatal = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plugging_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] virt_mkfs = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] volume_usage_poll_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] web = /usr/share/spice-html5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.769 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_concurrency.disable_process_locking = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_concurrency.lock_path = /var/lib/nova/tmp log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_metrics.metrics_buffer_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_metrics.metrics_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_metrics.metrics_process_name = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.770 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.auth_strategy = keystone log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.compute_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.dhcp_domain = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.enable_instance_password = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.glance_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.instance_list_cells_batch_fixed_size = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.771 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.instance_list_cells_batch_strategy = distributed log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.instance_list_per_project_cells = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.list_records_by_skipping_down_cells = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.local_metadata_per_cell = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.max_limit = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.metadata_cache_expiration = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.neutron_default_tenant_id = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.use_forwarded_for = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.772 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.use_neutron_default_nets = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_dynamic_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_dynamic_failure_fatal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_dynamic_read_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_dynamic_ssl_certfile = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_dynamic_targets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_jsonfile_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.773 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api.vendordata_providers = ['StaticJSON'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.backend = oslo_cache.dict log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.backend_argument = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.config_prefix = cache.oslo log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.dead_timeout = 60.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.debug_cache_backend = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.enable_retry_client = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.enable_socket_keepalive = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.774 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.expiration_time = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.hashclient_retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.hashclient_retry_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_dead_retry = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_password = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_pool_connection_get_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.775 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_pool_flush_on_reconnect = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_pool_maxsize = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_pool_unused_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_sasl_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_servers = ['localhost:11211'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_socket_timeout = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.memcache_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.proxies = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.776 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.retry_delay = 0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.socket_keepalive_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.socket_keepalive_idle = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.socket_keepalive_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.tls_allowed_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.tls_cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.tls_certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.777 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.tls_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cache.tls_keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.catalog_info = volumev3:cinderv3:internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.778 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.cross_az_attach = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.endpoint_template = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.779 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.os_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cinder.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.consecutive_build_service_disable_threshold = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.cpu_dedicated_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.cpu_shared_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.image_type_exclude_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.live_migration_wait_for_vif_plug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.780 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.max_concurrent_disk_ops = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.max_disk_devices_to_attach = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.packing_host_numa_cells_allocation_strategy = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.provider_config_location = /etc/nova/provider_config/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.resource_provider_association_refresh = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.shutdown_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] conductor.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.781 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] console.allowed_origins = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] console.ssl_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] console.ssl_minimum_version = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] consoleauth.token_ttl = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.782 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.783 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.service_type = accelerator log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] cyborg.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.784 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.785 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.786 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.787 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.788 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.789 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] api_database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] devices.enabled_mdev_types = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ephemeral_storage_encryption.cipher = aes-xts-plain64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ephemeral_storage_encryption.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ephemeral_storage_encryption.key_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.api_servers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.790 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.default_trusted_certificate_ids = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.enable_certificate_validation = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.791 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.enable_rbd_download = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.num_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.rbd_ceph_conf = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.792 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.rbd_pool = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.rbd_user = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.service_type = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.793 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.verify_glance_signatures = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] glance.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] guestfs.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.config_drive_cdrom = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.794 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.config_drive_inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.dynamic_memory_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.enable_instance_metrics_collection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.enable_remotefx = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.instances_path_share = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.iscsi_initiator_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.limit_cpu_features = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.mounted_disk_query_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.795 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.mounted_disk_query_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.power_state_check_timeframe = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.power_state_event_polling_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.qemu_img_cmd = qemu-img.exe log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.use_multipath_io = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.volume_attach_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.volume_attach_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.vswitch_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.796 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] hyperv.wait_soft_reboot_seconds = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] mks.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] mks.mksproxy_base_url = http://127.0.0.1:6090/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.manager_interval = 2400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.precache_concurrency = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.remove_unused_base_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.797 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.remove_unused_original_minimum_age_seconds = 86400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.remove_unused_resized_minimum_age_seconds = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] image_cache.subdirectory_name = _base log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.api_max_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.api_retry_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.798 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.799 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.partition_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.peer_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.serial_console_state_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.800 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.service_type = baremetal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ironic.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.801 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] key_manager.backend = barbican log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] key_manager.fixed_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.auth_endpoint = http://localhost/identity/v3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.barbican_api_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.barbican_endpoint = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.barbican_endpoint_type = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.barbican_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.802 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.number_of_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.retry_delay = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.send_service_user_token = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.803 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.verify_ssl = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican.verify_ssl_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.804 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] barbican_service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.approle_role_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.approle_secret_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.805 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.kv_mountpoint = secret log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.kv_version = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.root_token_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.806 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.use_ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vault.vault_url = http://127.0.0.1:8200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.807 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.808 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.service_type = identity log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.809 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] keystone.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.connection_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_mode = host-model log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_model_extra_flags = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_models = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_power_governor_high = performance log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_power_governor_low = powersave log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.810 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_power_management = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.cpu_power_management_strategy = cpu_state log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.device_detach_attempts = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.device_detach_timeout = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.disk_cachemodes = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.disk_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.enabled_perf_events = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.file_backed_memory = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.811 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.gid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.hw_disk_discard = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.hw_machine_type = ['x86_64=q35'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_rbd_ceph_conf = /etc/ceph/ceph.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_rbd_glance_copy_poll_interval = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_rbd_glance_copy_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_rbd_glance_store_name = default_backend log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.812 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_rbd_pool = vms log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_type = rbd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.images_volume_group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.inject_key = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.inject_partition = -2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.iscsi_iface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.iser_use_multipath = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.813 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_bandwidth = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_completion_timeout = 800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_downtime = 500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_downtime_delay = 75 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_downtime_steps = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_inbound_addr = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_permit_auto_converge = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_permit_post_copy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.814 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_scheme = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_timeout_action = force_complete log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_tunnelled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 WARNING oslo_config.cfg [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Deprecated: Option "live_migration_uri" from group "libvirt" is deprecated for removal ( Nov 27 04:34:34 localhost nova_compute[232618]: live_migration_uri is deprecated for removal in favor of two other options that Nov 27 04:34:34 localhost nova_compute[232618]: allow to change live migration scheme and target URI: ``live_migration_scheme`` Nov 27 04:34:34 localhost nova_compute[232618]: and ``live_migration_inbound_addr`` respectively. Nov 27 04:34:34 localhost nova_compute[232618]: ). Its value may be silently ignored in the future.#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_uri = qemu+ssh://nova@%s/system?keyfile=/var/lib/nova/.ssh/ssh-privatekey log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.live_migration_with_native_tls = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.815 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.max_queues = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.mem_stats_period_seconds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.nfs_mount_options = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.nfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_aoe_discover_tries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_iser_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_memory_encrypted_guests = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.816 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_nvme_discover_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_pcie_ports = 24 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.num_volume_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.pmem_namespaces = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.quobyte_client_cfg = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.quobyte_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rbd_destroy_volume_retries = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.817 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rbd_destroy_volume_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rbd_secret_uuid = e83f3b0c-4090-52df-95d4-ad9be8516692 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rbd_user = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.realtime_scheduler_priority = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.remote_filesystem_transport = ssh log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rescue_image_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rescue_kernel_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.818 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rescue_ramdisk_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rng_dev_path = /dev/urandom log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.rx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.smbfs_mount_options = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.smbfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.snapshot_compression = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.snapshot_image_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.819 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.snapshots_directory = /var/lib/nova/instances/snapshots log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.sparse_logical_volumes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.swtpm_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.swtpm_group = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.swtpm_user = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.sysinfo_serial = unique log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.tx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.820 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.uid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.use_virtio_for_bridges = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.virt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.volume_clear = zero log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.volume_clear_size = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.volume_use_multipath = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_cache_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.821 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_mount_group = qemu log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_mount_opts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_mount_perms = 0770 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.vzstorage_mount_user = stack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] libvirt.wait_soft_reboot_seconds = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.822 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.default_floating_pool = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.823 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.extension_sync_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.metadata_proxy_shared_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.ovs_bridge = br-int log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.824 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.physnets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.service_metadata_proxy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.service_type = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.825 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] neutron.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] notifications.bdms_in_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] notifications.default_level = INFO log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] notifications.notification_format = unversioned log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] notifications.notify_on_state_change = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.826 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] notifications.versioned_notifications_topics = ['versioned_notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] pci.alias = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] pci.device_spec = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] pci.report_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.827 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.828 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.829 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.project_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.project_name = service log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.service_type = placement log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.830 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.system_scope = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.831 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] placement.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.cores = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.count_usage_from_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.driver = nova.quota.DbQuotaDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.injected_file_content_bytes = 10240 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.injected_file_path_length = 255 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.832 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.injected_files = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.instances = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.key_pairs = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.metadata_items = 128 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.ram = 51200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.recheck_quota = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.server_group_members = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] quota.server_groups = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.833 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rdp.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] rdp.html5_proxy_base_url = http://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.discover_hosts_in_cells_interval = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.enable_isolated_aggregate_filtering = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.image_metadata_prefilter = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.limit_tenants_to_placement_aggregate = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.max_attempts = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.834 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.max_placement_results = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.placement_aggregate_required_for_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.query_placement_for_availability_zone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.query_placement_for_image_type_support = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.query_placement_for_routed_network_aggregates = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] scheduler.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.aggregate_image_properties_isolation_namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.835 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.aggregate_image_properties_isolation_separator = . log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.build_failure_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.cpu_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.disk_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.836 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.host_subset_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.image_properties_default_architecture = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.io_ops_weight_multiplier = -1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.isolated_hosts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.isolated_images = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.max_instances_per_host = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.max_io_ops_per_host = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.pci_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.837 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.pci_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.ram_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.shuffle_best_same_weighed_hosts = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.soft_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.track_instance_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.838 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metrics.required = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metrics.weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metrics.weight_of_unavailable = -10000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] metrics.weight_setting = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.base_url = ws://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.port_range = 10000:20000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.839 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.serialproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] serial_console.serialproxy_port = 6083 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.840 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.send_service_user_token = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.agent_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.841 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.html5proxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.html5proxy_port = 6082 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.image_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.jpeg_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.playback_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.server_listen = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.842 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.server_proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.streaming_mode = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] spice.zlib_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] upgrade_levels.baseapi = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] upgrade_levels.cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] upgrade_levels.compute = auto log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] upgrade_levels.conductor = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] upgrade_levels.scheduler = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.843 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.844 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vendordata_dynamic_auth.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.cache_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.cluster_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.connection_pool_size = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.console_delay_seconds = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.845 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.datastore_regex = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.host_ip = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.host_username = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.integration_bridge = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.maximum_objects = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.846 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.pbm_default_policy = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.pbm_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.pbm_wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.serial_log_dir = /opt/vmware/vspc log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.serial_port_proxy_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.serial_port_service_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.use_linked_clone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.847 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.vnc_keymap = en-us log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.vnc_port = 5900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vmware.vnc_port_total = 10000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.auth_schemes = ['none'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.novncproxy_base_url = http://nova-novncproxy-cell1-public-openstack.apps-crc.testing/vnc_lite.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.novncproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.848 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.novncproxy_port = 6080 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.server_listen = ::0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.server_proxyclient_address = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.vencrypt_ca_certs = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.vencrypt_client_cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vnc.vencrypt_client_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_compute_service_check_for_ffu = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.849 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_deep_image_inspection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_fallback_pcpu_query = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_group_policy_check_upcall = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_libvirt_livesnapshot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.disable_rootwrap = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.enable_numa_live_migration = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.enable_qemu_monitor_announce_self = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.850 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.handle_virt_lifecycle_events = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.libvirt_disable_apic = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.never_download_image_if_on_rbd = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.qemu_monitor_announce_self_count = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.qemu_monitor_announce_self_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.reserve_disk_resource_for_image_cache = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.skip_cpu_compare_at_startup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.851 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.skip_cpu_compare_on_dest = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.skip_hypervisor_version_check_on_lm = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.skip_reserve_in_use_ironic_nodes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.unified_limits_count_pcpu_as_vcpu = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.api_paste_config = api-paste.ini log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.client_socket_timeout = 900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.default_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.852 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.keep_alive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.max_header_line = 16384 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.secure_proxy_ssl_header = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.ssl_ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.ssl_cert_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.ssl_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.tcp_keepidle = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.853 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] zvm.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] zvm.cloud_connector_url = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] zvm.image_tmp_path = /var/lib/nova/images log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] zvm.reachable_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.enforce_new_defaults = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost python3.9[232744]: ansible-containers.podman.podman_container Invoked with name=nova_compute_init state=started executable=podman detach=True debug=False force_restart=False force_delete=True generate_systemd={} image_strict=False recreate=False image=None annotation=None arch=None attach=None authfile=None blkio_weight=None blkio_weight_device=None cap_add=None cap_drop=None cgroup_conf=None cgroup_parent=None cgroupns=None cgroups=None chrootdirs=None cidfile=None cmd_args=None conmon_pidfile=None command=None cpu_period=None cpu_quota=None cpu_rt_period=None cpu_rt_runtime=None cpu_shares=None cpus=None cpuset_cpus=None cpuset_mems=None decryption_key=None delete_depend=None delete_time=None delete_volumes=None detach_keys=None device=None device_cgroup_rule=None device_read_bps=None device_read_iops=None device_write_bps=None device_write_iops=None dns=None dns_option=None dns_search=None entrypoint=None env=None env_file=None env_host=None env_merge=None etc_hosts=None expose=None gidmap=None gpus=None group_add=None group_entry=None healthcheck=None healthcheck_interval=None healthcheck_retries=None healthcheck_start_period=None health_startup_cmd=None health_startup_interval=None health_startup_retries=None health_startup_success=None health_startup_timeout=None healthcheck_timeout=None healthcheck_failure_action=None hooks_dir=None hostname=None hostuser=None http_proxy=None image_volume=None init=None init_ctr=None init_path=None interactive=None ip=None ip6=None ipc=None kernel_memory=None label=None label_file=None log_driver=None log_level=None log_opt=None mac_address=None memory=None memory_reservation=None memory_swap=None memory_swappiness=None mount=None network=None network_aliases=None no_healthcheck=None no_hosts=None oom_kill_disable=None oom_score_adj=None os=None passwd=None passwd_entry=None personality=None pid=None pid_file=None pids_limit=None platform=None pod=None pod_id_file=None preserve_fd=None preserve_fds=None privileged=None publish=None publish_all=None pull=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None rdt_class=None read_only=None read_only_tmpfs=None requires=None restart_policy=None restart_time=None retry=None retry_delay=None rm=None rmi=None rootfs=None seccomp_policy=None secrets=NOT_LOGGING_PARAMETER sdnotify=None security_opt=None shm_size=None shm_size_systemd=None sig_proxy=None stop_signal=None stop_timeout=None stop_time=None subgidname=None subuidname=None sysctl=None systemd=None timeout=None timezone=None tls_verify=None tmpfs=None tty=None uidmap=None ulimit=None umask=None unsetenv=None unsetenv_all=None user=None userns=None uts=None variant=None volume=None volumes_from=None workdir=None Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.enforce_scope = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.policy_default_rule = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.854 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.policy_dirs = ['policy.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.policy_file = policy.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.remote_content_type = application/x-www-form-urlencoded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.remote_ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.remote_ssl_client_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.remote_ssl_client_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_policy.remote_ssl_verify_server_crt = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.855 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_versionedobjects.fatal_exception_format_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] remote_debug.host = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] remote_debug.port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.amqp_durable_queues = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.856 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.857 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.858 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_queue = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.859 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_notifications.driver = ['noop'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.860 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.861 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.endpoint_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.862 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.project_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.863 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.project_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.service_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.system_scope = all log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.864 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.valid_interfaces = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.865 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_limit.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_reports.file_event_handler = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_reports.file_event_handler_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] oslo_reports.log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.capabilities = [12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.866 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_linux_bridge_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.capabilities = [12, 1] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] vif_plug_ovs_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.867 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.flat_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.forward_bridge_interface = ['all'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.iptables_bottom_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.iptables_drop_action = DROP log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.iptables_top_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.use_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.868 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_linux_bridge.vlan_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.isolate_vif = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.ovs_vsctl_timeout = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.ovsdb_interface = native log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_vif_ovs.per_port_bridge = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_brick.lock_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.869 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_brick.wait_mpath_device_attempts = 4 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] os_brick.wait_mpath_device_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.capabilities = [21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.logger_name = os_brick.privileged log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] privsep_osbrick.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.870 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] nova_sys_admin.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.871 232622 DEBUG oslo_service.service [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.873 232622 INFO nova.service [-] Starting compute node (version 27.5.2-0.20250829104910.6f8decf.el9)#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.887 232622 INFO nova.virt.node [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.887 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Starting native event thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:492#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.887 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Starting green dispatch thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:498#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.888 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Starting connection event dispatch thread initialize /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:620#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.888 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Connecting to libvirt: qemu:///system _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:503#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.898 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Registering for lifecycle events _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:509#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.900 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Registering for connection events: _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:530#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.901 232622 INFO nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Connection event '1' reason 'None'#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.906 232622 INFO nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Libvirt host capabilities Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 8a8a8082-6126-4917-bb42-c4150a0ac6a5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: x86_64 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v4 Nov 27 04:34:34 localhost nova_compute[232618]: AMD Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: tcp Nov 27 04:34:34 localhost nova_compute[232618]: rdma Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 16116612 Nov 27 04:34:34 localhost nova_compute[232618]: 4029153 Nov 27 04:34:34 localhost nova_compute[232618]: 0 Nov 27 04:34:34 localhost nova_compute[232618]: 0 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: selinux Nov 27 04:34:34 localhost nova_compute[232618]: 0 Nov 27 04:34:34 localhost nova_compute[232618]: system_u:system_r:svirt_t:s0 Nov 27 04:34:34 localhost nova_compute[232618]: system_u:system_r:svirt_tcg_t:s0 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: dac Nov 27 04:34:34 localhost nova_compute[232618]: 0 Nov 27 04:34:34 localhost nova_compute[232618]: +107:+107 Nov 27 04:34:34 localhost nova_compute[232618]: +107:+107 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: hvm Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 32 Nov 27 04:34:34 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:34 localhost nova_compute[232618]: pc-i440fx-rhel7.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.8.0 Nov 27 04:34:34 localhost nova_compute[232618]: q35 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.4.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.5.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.3.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel7.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.4.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.2.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.2.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.0.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.0.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.1.0 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: hvm Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 64 Nov 27 04:34:34 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:34 localhost nova_compute[232618]: pc-i440fx-rhel7.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.8.0 Nov 27 04:34:34 localhost nova_compute[232618]: q35 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.4.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.5.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.3.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel7.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.4.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.2.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.2.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.0.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.0.0 Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel8.1.0 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: #033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.913 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Getting domain capabilities for i686 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.916 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=pc: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:34 localhost nova_compute[232618]: kvm Nov 27 04:34:34 localhost nova_compute[232618]: pc-i440fx-rhel7.6.0 Nov 27 04:34:34 localhost nova_compute[232618]: i686 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: rom Nov 27 04:34:34 localhost nova_compute[232618]: pflash Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: yes Nov 27 04:34:34 localhost nova_compute[232618]: no Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: no Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: on Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: on Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:34 localhost nova_compute[232618]: AMD Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 486 Nov 27 04:34:34 localhost nova_compute[232618]: 486-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Conroe Nov 27 04:34:34 localhost nova_compute[232618]: Conroe-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Genoa Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Genoa-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-IBPB Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v4 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v1 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v2 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v6 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server-v7 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: IvyBridge Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: IvyBridge-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: IvyBridge-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: IvyBridge-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: KnightsMill Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: KnightsMill-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nehalem Nov 27 04:34:34 localhost nova_compute[232618]: Nehalem-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nehalem-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nehalem-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G1 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G1-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G2 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G2-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G3 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G3-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G4-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Opteron_G5-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Penryn Nov 27 04:34:34 localhost nova_compute[232618]: Penryn-v1 Nov 27 04:34:34 localhost nova_compute[232618]: SandyBridge Nov 27 04:34:34 localhost nova_compute[232618]: SandyBridge-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: SandyBridge-v1 Nov 27 04:34:34 localhost nova_compute[232618]: SandyBridge-v2 Nov 27 04:34:34 localhost nova_compute[232618]: SapphireRapids Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: SapphireRapids-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: SapphireRapids-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: SapphireRapids-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: SierraForest Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: SierraForest-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Client-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Skylake-Server-v5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Snowridge Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Snowridge-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Snowridge-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Snowridge-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Snowridge-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Westmere Nov 27 04:34:34 localhost nova_compute[232618]: Westmere-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Westmere-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Westmere-v2 Nov 27 04:34:34 localhost nova_compute[232618]: athlon Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: athlon-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: core2duo Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: core2duo-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: coreduo Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: coreduo-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: kvm32 Nov 27 04:34:34 localhost nova_compute[232618]: kvm32-v1 Nov 27 04:34:34 localhost nova_compute[232618]: kvm64 Nov 27 04:34:34 localhost nova_compute[232618]: kvm64-v1 Nov 27 04:34:34 localhost nova_compute[232618]: n270 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: n270-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: pentium Nov 27 04:34:34 localhost nova_compute[232618]: pentium-v1 Nov 27 04:34:34 localhost nova_compute[232618]: pentium2 Nov 27 04:34:34 localhost nova_compute[232618]: pentium2-v1 Nov 27 04:34:34 localhost nova_compute[232618]: pentium3 Nov 27 04:34:34 localhost nova_compute[232618]: pentium3-v1 Nov 27 04:34:34 localhost nova_compute[232618]: phenom Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: phenom-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: qemu32 Nov 27 04:34:34 localhost nova_compute[232618]: qemu32-v1 Nov 27 04:34:34 localhost nova_compute[232618]: qemu64 Nov 27 04:34:34 localhost nova_compute[232618]: qemu64-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: file Nov 27 04:34:34 localhost nova_compute[232618]: anonymous Nov 27 04:34:34 localhost nova_compute[232618]: memfd Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: disk Nov 27 04:34:34 localhost nova_compute[232618]: cdrom Nov 27 04:34:34 localhost nova_compute[232618]: floppy Nov 27 04:34:34 localhost nova_compute[232618]: lun Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: ide Nov 27 04:34:34 localhost nova_compute[232618]: fdc Nov 27 04:34:34 localhost nova_compute[232618]: scsi Nov 27 04:34:34 localhost nova_compute[232618]: virtio Nov 27 04:34:34 localhost nova_compute[232618]: usb Nov 27 04:34:34 localhost nova_compute[232618]: sata Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: virtio Nov 27 04:34:34 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:34 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: vnc Nov 27 04:34:34 localhost nova_compute[232618]: egl-headless Nov 27 04:34:34 localhost nova_compute[232618]: dbus Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: subsystem Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: default Nov 27 04:34:34 localhost nova_compute[232618]: mandatory Nov 27 04:34:34 localhost nova_compute[232618]: requisite Nov 27 04:34:34 localhost nova_compute[232618]: optional Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: usb Nov 27 04:34:34 localhost nova_compute[232618]: pci Nov 27 04:34:34 localhost nova_compute[232618]: scsi Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: virtio Nov 27 04:34:34 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:34 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: random Nov 27 04:34:34 localhost nova_compute[232618]: egd Nov 27 04:34:34 localhost nova_compute[232618]: builtin Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: path Nov 27 04:34:34 localhost nova_compute[232618]: handle Nov 27 04:34:34 localhost nova_compute[232618]: virtiofs Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: tpm-tis Nov 27 04:34:34 localhost nova_compute[232618]: tpm-crb Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: emulator Nov 27 04:34:34 localhost nova_compute[232618]: external Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 2.0 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: usb Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: pty Nov 27 04:34:34 localhost nova_compute[232618]: unix Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: qemu Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: builtin Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: default Nov 27 04:34:34 localhost nova_compute[232618]: passt Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: isa Nov 27 04:34:34 localhost nova_compute[232618]: hyperv Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: null Nov 27 04:34:34 localhost nova_compute[232618]: vc Nov 27 04:34:34 localhost nova_compute[232618]: pty Nov 27 04:34:34 localhost nova_compute[232618]: dev Nov 27 04:34:34 localhost nova_compute[232618]: file Nov 27 04:34:34 localhost nova_compute[232618]: pipe Nov 27 04:34:34 localhost nova_compute[232618]: stdio Nov 27 04:34:34 localhost nova_compute[232618]: udp Nov 27 04:34:34 localhost nova_compute[232618]: tcp Nov 27 04:34:34 localhost nova_compute[232618]: unix Nov 27 04:34:34 localhost nova_compute[232618]: qemu-vdagent Nov 27 04:34:34 localhost nova_compute[232618]: dbus Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: relaxed Nov 27 04:34:34 localhost nova_compute[232618]: vapic Nov 27 04:34:34 localhost nova_compute[232618]: spinlocks Nov 27 04:34:34 localhost nova_compute[232618]: vpindex Nov 27 04:34:34 localhost nova_compute[232618]: runtime Nov 27 04:34:34 localhost nova_compute[232618]: synic Nov 27 04:34:34 localhost nova_compute[232618]: stimer Nov 27 04:34:34 localhost nova_compute[232618]: reset Nov 27 04:34:34 localhost nova_compute[232618]: vendor_id Nov 27 04:34:34 localhost nova_compute[232618]: frequencies Nov 27 04:34:34 localhost nova_compute[232618]: reenlightenment Nov 27 04:34:34 localhost nova_compute[232618]: tlbflush Nov 27 04:34:34 localhost nova_compute[232618]: ipi Nov 27 04:34:34 localhost nova_compute[232618]: avic Nov 27 04:34:34 localhost nova_compute[232618]: emsr_bitmap Nov 27 04:34:34 localhost nova_compute[232618]: xmm_input Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 4095 Nov 27 04:34:34 localhost nova_compute[232618]: on Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: Linux KVM Hv Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: tdx Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.919 232622 DEBUG nova.virt.libvirt.volume.mount [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Initialising _HostMountState generation 0 host_up /usr/lib/python3.9/site-packages/nova/virt/libvirt/volume/mount.py:130#033[00m Nov 27 04:34:34 localhost nova_compute[232618]: 2025-11-27 09:34:34.922 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=q35: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:34 localhost nova_compute[232618]: kvm Nov 27 04:34:34 localhost nova_compute[232618]: pc-q35-rhel9.8.0 Nov 27 04:34:34 localhost nova_compute[232618]: i686 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: rom Nov 27 04:34:34 localhost nova_compute[232618]: pflash Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: yes Nov 27 04:34:34 localhost nova_compute[232618]: no Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: no Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: on Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: on Nov 27 04:34:34 localhost nova_compute[232618]: off Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:34 localhost nova_compute[232618]: AMD Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: 486 Nov 27 04:34:34 localhost nova_compute[232618]: 486-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Broadwell-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cascadelake-Server-v5 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Conroe Nov 27 04:34:34 localhost nova_compute[232618]: Conroe-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Cooperlake-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Denverton-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Dhyana-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Genoa Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Genoa-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-IBPB Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Milan-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-Rome-v4 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v1 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v2 Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: EPYC-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: GraniteRapids-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-noTSX Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-noTSX-IBRS Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v1 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v2 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v3 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Haswell-v4 Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Icelake-Server Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:34 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v6 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v7 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Penryn Nov 27 04:34:35 localhost nova_compute[232618]: Penryn-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Westmere Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v2 Nov 27 04:34:35 localhost nova_compute[232618]: athlon Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: athlon-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: kvm32 Nov 27 04:34:35 localhost nova_compute[232618]: kvm32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: n270 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: n270-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pentium Nov 27 04:34:35 localhost nova_compute[232618]: pentium-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: phenom Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: phenom-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu32 Nov 27 04:34:35 localhost nova_compute[232618]: qemu32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: anonymous Nov 27 04:34:35 localhost nova_compute[232618]: memfd Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: disk Nov 27 04:34:35 localhost nova_compute[232618]: cdrom Nov 27 04:34:35 localhost nova_compute[232618]: floppy Nov 27 04:34:35 localhost nova_compute[232618]: lun Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: fdc Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: sata Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: vnc Nov 27 04:34:35 localhost nova_compute[232618]: egl-headless Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: subsystem Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: mandatory Nov 27 04:34:35 localhost nova_compute[232618]: requisite Nov 27 04:34:35 localhost nova_compute[232618]: optional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: pci Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: random Nov 27 04:34:35 localhost nova_compute[232618]: egd Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: path Nov 27 04:34:35 localhost nova_compute[232618]: handle Nov 27 04:34:35 localhost nova_compute[232618]: virtiofs Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tpm-tis Nov 27 04:34:35 localhost nova_compute[232618]: tpm-crb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: emulator Nov 27 04:34:35 localhost nova_compute[232618]: external Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 2.0 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: passt Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: isa Nov 27 04:34:35 localhost nova_compute[232618]: hyperv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: null Nov 27 04:34:35 localhost nova_compute[232618]: vc Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: dev Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: pipe Nov 27 04:34:35 localhost nova_compute[232618]: stdio Nov 27 04:34:35 localhost nova_compute[232618]: udp Nov 27 04:34:35 localhost nova_compute[232618]: tcp Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: qemu-vdagent Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: relaxed Nov 27 04:34:35 localhost nova_compute[232618]: vapic Nov 27 04:34:35 localhost nova_compute[232618]: spinlocks Nov 27 04:34:35 localhost nova_compute[232618]: vpindex Nov 27 04:34:35 localhost nova_compute[232618]: runtime Nov 27 04:34:35 localhost nova_compute[232618]: synic Nov 27 04:34:35 localhost nova_compute[232618]: stimer Nov 27 04:34:35 localhost nova_compute[232618]: reset Nov 27 04:34:35 localhost nova_compute[232618]: vendor_id Nov 27 04:34:35 localhost nova_compute[232618]: frequencies Nov 27 04:34:35 localhost nova_compute[232618]: reenlightenment Nov 27 04:34:35 localhost nova_compute[232618]: tlbflush Nov 27 04:34:35 localhost nova_compute[232618]: ipi Nov 27 04:34:35 localhost nova_compute[232618]: avic Nov 27 04:34:35 localhost nova_compute[232618]: emsr_bitmap Nov 27 04:34:35 localhost nova_compute[232618]: xmm_input Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 4095 Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Linux KVM Hv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tdx Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:34.952 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Getting domain capabilities for x86_64 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:34.958 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=pc: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:35 localhost nova_compute[232618]: kvm Nov 27 04:34:35 localhost nova_compute[232618]: pc-i440fx-rhel7.6.0 Nov 27 04:34:35 localhost nova_compute[232618]: x86_64 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: rom Nov 27 04:34:35 localhost nova_compute[232618]: pflash Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: yes Nov 27 04:34:35 localhost nova_compute[232618]: no Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: no Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:35 localhost nova_compute[232618]: AMD Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 486 Nov 27 04:34:35 localhost nova_compute[232618]: 486-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Conroe Nov 27 04:34:35 localhost nova_compute[232618]: Conroe-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Genoa Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Genoa-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-IBPB Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v4 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v1 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v2 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v6 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v7 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Penryn Nov 27 04:34:35 localhost nova_compute[232618]: Penryn-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Westmere Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v2 Nov 27 04:34:35 localhost nova_compute[232618]: athlon Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: athlon-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: kvm32 Nov 27 04:34:35 localhost nova_compute[232618]: kvm32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: n270 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: n270-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pentium Nov 27 04:34:35 localhost nova_compute[232618]: pentium-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: phenom Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: phenom-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu32 Nov 27 04:34:35 localhost nova_compute[232618]: qemu32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: anonymous Nov 27 04:34:35 localhost nova_compute[232618]: memfd Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: disk Nov 27 04:34:35 localhost nova_compute[232618]: cdrom Nov 27 04:34:35 localhost nova_compute[232618]: floppy Nov 27 04:34:35 localhost nova_compute[232618]: lun Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: ide Nov 27 04:34:35 localhost nova_compute[232618]: fdc Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: sata Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: vnc Nov 27 04:34:35 localhost nova_compute[232618]: egl-headless Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: subsystem Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: mandatory Nov 27 04:34:35 localhost nova_compute[232618]: requisite Nov 27 04:34:35 localhost nova_compute[232618]: optional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: pci Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: random Nov 27 04:34:35 localhost nova_compute[232618]: egd Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: path Nov 27 04:34:35 localhost nova_compute[232618]: handle Nov 27 04:34:35 localhost nova_compute[232618]: virtiofs Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tpm-tis Nov 27 04:34:35 localhost nova_compute[232618]: tpm-crb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: emulator Nov 27 04:34:35 localhost nova_compute[232618]: external Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 2.0 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: passt Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: isa Nov 27 04:34:35 localhost nova_compute[232618]: hyperv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: null Nov 27 04:34:35 localhost nova_compute[232618]: vc Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: dev Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: pipe Nov 27 04:34:35 localhost nova_compute[232618]: stdio Nov 27 04:34:35 localhost nova_compute[232618]: udp Nov 27 04:34:35 localhost nova_compute[232618]: tcp Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: qemu-vdagent Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: relaxed Nov 27 04:34:35 localhost nova_compute[232618]: vapic Nov 27 04:34:35 localhost nova_compute[232618]: spinlocks Nov 27 04:34:35 localhost nova_compute[232618]: vpindex Nov 27 04:34:35 localhost nova_compute[232618]: runtime Nov 27 04:34:35 localhost nova_compute[232618]: synic Nov 27 04:34:35 localhost nova_compute[232618]: stimer Nov 27 04:34:35 localhost nova_compute[232618]: reset Nov 27 04:34:35 localhost nova_compute[232618]: vendor_id Nov 27 04:34:35 localhost nova_compute[232618]: frequencies Nov 27 04:34:35 localhost nova_compute[232618]: reenlightenment Nov 27 04:34:35 localhost nova_compute[232618]: tlbflush Nov 27 04:34:35 localhost nova_compute[232618]: ipi Nov 27 04:34:35 localhost nova_compute[232618]: avic Nov 27 04:34:35 localhost nova_compute[232618]: emsr_bitmap Nov 27 04:34:35 localhost nova_compute[232618]: xmm_input Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 4095 Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Linux KVM Hv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tdx Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.011 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=q35: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: /usr/libexec/qemu-kvm Nov 27 04:34:35 localhost nova_compute[232618]: kvm Nov 27 04:34:35 localhost nova_compute[232618]: pc-q35-rhel9.8.0 Nov 27 04:34:35 localhost nova_compute[232618]: x86_64 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: efi Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: /usr/share/edk2/ovmf/OVMF_CODE.secboot.fd Nov 27 04:34:35 localhost nova_compute[232618]: /usr/share/edk2/ovmf/OVMF_CODE.fd Nov 27 04:34:35 localhost nova_compute[232618]: /usr/share/edk2/ovmf/OVMF.amdsev.fd Nov 27 04:34:35 localhost nova_compute[232618]: /usr/share/edk2/ovmf/OVMF.inteltdx.secboot.fd Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: rom Nov 27 04:34:35 localhost nova_compute[232618]: pflash Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: yes Nov 27 04:34:35 localhost nova_compute[232618]: no Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: yes Nov 27 04:34:35 localhost nova_compute[232618]: no Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:35 localhost nova_compute[232618]: AMD Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 486 Nov 27 04:34:35 localhost nova_compute[232618]: 486-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Broadwell-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost systemd[1]: Started libpod-conmon-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope. Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/usr/sbin/nova_statedir_ownership.py supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cascadelake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/var/lib/_nova_secontext supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Conroe Nov 27 04:34:35 localhost nova_compute[232618]: Conroe-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Cooperlake-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Denverton-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Dhyana-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Genoa Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Genoa-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-IBPB Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Milan-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-Rome-v4 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v1 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v2 Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: EPYC-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: GraniteRapids-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost systemd[1]: Started libcrun container. Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Haswell-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-noTSX Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v6 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Icelake-Server-v7 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: IvyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: KnightsMill-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nehalem-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G1-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G4-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Opteron_G5-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Penryn Nov 27 04:34:35 localhost nova_compute[232618]: Penryn-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: SandyBridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost podman[232790]: 2025-11-27 09:34:35.119819902 +0000 UTC m=+0.119045450 container init 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, container_name=nova_compute_init, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SapphireRapids-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: SierraForest-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Client-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-noTSX-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Skylake-Server-v5 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v2 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v3 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Snowridge-v4 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Westmere Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-IBRS Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Westmere-v2 Nov 27 04:34:35 localhost nova_compute[232618]: athlon Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: athlon-v1 Nov 27 04:34:35 localhost podman[232790]: 2025-11-27 09:34:35.135594932 +0000 UTC m=+0.134820490 container start 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, container_name=nova_compute_init, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: core2duo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: coreduo-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: kvm32 Nov 27 04:34:35 localhost nova_compute[232618]: kvm32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64 Nov 27 04:34:35 localhost nova_compute[232618]: kvm64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: n270 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: n270-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pentium Nov 27 04:34:35 localhost nova_compute[232618]: pentium-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2 Nov 27 04:34:35 localhost nova_compute[232618]: pentium2-v1 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3 Nov 27 04:34:35 localhost nova_compute[232618]: pentium3-v1 Nov 27 04:34:35 localhost nova_compute[232618]: phenom Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: phenom-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu32 Nov 27 04:34:35 localhost nova_compute[232618]: qemu32-v1 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64 Nov 27 04:34:35 localhost nova_compute[232618]: qemu64-v1 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: anonymous Nov 27 04:34:35 localhost nova_compute[232618]: memfd Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: disk Nov 27 04:34:35 localhost nova_compute[232618]: cdrom Nov 27 04:34:35 localhost nova_compute[232618]: floppy Nov 27 04:34:35 localhost nova_compute[232618]: lun Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: fdc Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: sata Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: vnc Nov 27 04:34:35 localhost nova_compute[232618]: egl-headless Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: subsystem Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: mandatory Nov 27 04:34:35 localhost nova_compute[232618]: requisite Nov 27 04:34:35 localhost nova_compute[232618]: optional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: pci Nov 27 04:34:35 localhost nova_compute[232618]: scsi Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: virtio Nov 27 04:34:35 localhost nova_compute[232618]: virtio-transitional Nov 27 04:34:35 localhost nova_compute[232618]: virtio-non-transitional Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: random Nov 27 04:34:35 localhost nova_compute[232618]: egd Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: path Nov 27 04:34:35 localhost nova_compute[232618]: handle Nov 27 04:34:35 localhost nova_compute[232618]: virtiofs Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tpm-tis Nov 27 04:34:35 localhost nova_compute[232618]: tpm-crb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: emulator Nov 27 04:34:35 localhost nova_compute[232618]: external Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost python3.9[232744]: ansible-containers.podman.podman_container PODMAN-CONTAINER-DEBUG: podman start nova_compute_init Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 2.0 Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: usb Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: qemu Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: builtin Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: default Nov 27 04:34:35 localhost nova_compute[232618]: passt Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: isa Nov 27 04:34:35 localhost nova_compute[232618]: hyperv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: null Nov 27 04:34:35 localhost nova_compute[232618]: vc Nov 27 04:34:35 localhost nova_compute[232618]: pty Nov 27 04:34:35 localhost nova_compute[232618]: dev Nov 27 04:34:35 localhost nova_compute[232618]: file Nov 27 04:34:35 localhost nova_compute[232618]: pipe Nov 27 04:34:35 localhost nova_compute[232618]: stdio Nov 27 04:34:35 localhost nova_compute[232618]: udp Nov 27 04:34:35 localhost nova_compute[232618]: tcp Nov 27 04:34:35 localhost nova_compute[232618]: unix Nov 27 04:34:35 localhost nova_compute[232618]: qemu-vdagent Nov 27 04:34:35 localhost nova_compute[232618]: dbus Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: relaxed Nov 27 04:34:35 localhost nova_compute[232618]: vapic Nov 27 04:34:35 localhost nova_compute[232618]: spinlocks Nov 27 04:34:35 localhost nova_compute[232618]: vpindex Nov 27 04:34:35 localhost nova_compute[232618]: runtime Nov 27 04:34:35 localhost nova_compute[232618]: synic Nov 27 04:34:35 localhost nova_compute[232618]: stimer Nov 27 04:34:35 localhost nova_compute[232618]: reset Nov 27 04:34:35 localhost nova_compute[232618]: vendor_id Nov 27 04:34:35 localhost nova_compute[232618]: frequencies Nov 27 04:34:35 localhost nova_compute[232618]: reenlightenment Nov 27 04:34:35 localhost nova_compute[232618]: tlbflush Nov 27 04:34:35 localhost nova_compute[232618]: ipi Nov 27 04:34:35 localhost nova_compute[232618]: avic Nov 27 04:34:35 localhost nova_compute[232618]: emsr_bitmap Nov 27 04:34:35 localhost nova_compute[232618]: xmm_input Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: 4095 Nov 27 04:34:35 localhost nova_compute[232618]: on Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: off Nov 27 04:34:35 localhost nova_compute[232618]: Linux KVM Hv Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: tdx Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: Nov 27 04:34:35 localhost nova_compute[232618]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.070 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.071 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.071 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.071 232622 INFO nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Secure Boot support detected#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.073 232622 INFO nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.073 232622 INFO nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.082 232622 DEBUG nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Enabling emulated TPM support _check_vtpm_support /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:1097#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.104 232622 INFO nova.virt.node [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.131 232622 DEBUG nova.compute.manager [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Verified node 5764ad4c-cf89-4d5f-a185-92be71f7b67f matches my host np0005537446.localdomain _check_for_host_rename /usr/lib/python3.9/site-packages/nova/compute/manager.py:1568#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.162 232622 DEBUG nova.compute.manager [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.166 232622 DEBUG nova.virt.libvirt.vif [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=,hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=,node='np0005537446.localdomain',numa_topology=None,old_flavor=,os_type=None,pci_devices=,pci_requests=,power_state=1,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata=,tags=,task_state=None,terminated_at=None,trusted_certs=,updated_at=2025-11-27T08:33:11Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"dhcp_server": "192.168.0.1"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system"}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.167 232622 DEBUG nova.network.os_vif_util [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"dhcp_server": "192.168.0.1"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system"}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.167 232622 DEBUG nova.network.os_vif_util [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Converted object VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.168 232622 DEBUG os_vif [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Plugging vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Applying nova statedir ownership Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Target ownership for /var/lib/nova: 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 1000 gid: 1000 path: /var/lib/nova/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Changing ownership of /var/lib/nova from 1000:1000 to 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 1000 gid: 1000 path: /var/lib/nova/instances/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Changing ownership of /var/lib/nova/instances from 1000:1000 to 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201 already 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201 to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 0 gid: 0 path: /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201/console.log Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/_base already 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/_base to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/454c8fd44a4104b436504db7d917f75244a9e94a Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/ephemeral_1_0706d66 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/locks already 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/locks to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/nova-454c8fd44a4104b436504db7d917f75244a9e94a Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/nova-ephemeral_1_0706d66 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 0 gid: 0 path: /var/lib/nova/delay-nova-compute Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/.ssh already 42436:42436 Nov 27 04:34:35 localhost systemd[1]: libpod-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope: Deactivated successfully. Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.ssh to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/config Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/.cache already 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.cache to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/ Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Ownership of /var/lib/nova/.cache/python-entrypoints already 42436:42436 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.cache/python-entrypoints to system_u:object_r:container_file_t:s0 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/b234715fc878456b41e32c4fbc669b417044dbe6c6684bbc9059e5c93396ffea Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/469bc4441baff9216df986857f9ff45dbf25965a8d2f755a6449ac2645cb7191 Nov 27 04:34:35 localhost nova_compute_init[232809]: INFO:nova_statedir:Nova statedir ownership complete Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.245 232622 DEBUG ovsdbapp.backend.ovs_idl [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Created schema index Interface.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.246 232622 DEBUG ovsdbapp.backend.ovs_idl [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Created schema index Port.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.246 232622 DEBUG ovsdbapp.backend.ovs_idl [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Created schema index Bridge.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.246 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] tcp:127.0.0.1:6640: entering CONNECTING _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.246 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [POLLOUT] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.247 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.247 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.248 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.250 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.260 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.261 232622 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.261 232622 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.262 232622 INFO oslo.privsep.daemon [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Running privsep helper: ['sudo', 'nova-rootwrap', '/etc/nova/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-compute.conf', '--config-dir', '/etc/nova/nova.conf.d', '--privsep_context', 'vif_plug_ovs.privsep.vif_plug', '--privsep_sock_path', '/tmp/tmpyz59o7_c/privsep.sock']#033[00m Nov 27 04:34:35 localhost podman[232829]: 2025-11-27 09:34:35.34390157 +0000 UTC m=+0.093912459 container died 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, container_name=nova_compute_init, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:34:35 localhost podman[232829]: 2025-11-27 09:34:35.394673637 +0000 UTC m=+0.144684496 container cleanup 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, container_name=nova_compute_init, io.buildah.version=1.41.3, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:34:35 localhost systemd[1]: libpod-conmon-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope: Deactivated successfully. Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.853 232622 INFO oslo.privsep.daemon [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.754 232866 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.761 232866 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.765 232866 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_DAC_OVERRIDE|CAP_NET_ADMIN/CAP_DAC_OVERRIDE|CAP_NET_ADMIN/none#033[00m Nov 27 04:34:35 localhost nova_compute[232618]: 2025-11-27 09:34:35.765 232866 INFO oslo.privsep.daemon [-] privsep daemon running as pid 232866#033[00m Nov 27 04:34:36 localhost systemd[1]: tmp-crun.AWCJ0G.mount: Deactivated successfully. Nov 27 04:34:36 localhost systemd[1]: var-lib-containers-storage-overlay-16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae-merged.mount: Deactivated successfully. Nov 27 04:34:36 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac-userdata-shm.mount: Deactivated successfully. Nov 27 04:34:36 localhost systemd[1]: session-54.scope: Deactivated successfully. Nov 27 04:34:36 localhost systemd[1]: session-54.scope: Consumed 2min 9.210s CPU time. Nov 27 04:34:36 localhost systemd-logind[761]: Session 54 logged out. Waiting for processes to exit. Nov 27 04:34:36 localhost systemd-logind[761]: Removed session 54. Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.129 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.129 232622 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tapa2718872-3b, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.130 232622 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tapa2718872-3b, col_values=(('external_ids', {'iface-id': 'a2718872-3b82-44ca-8d67-a0e516b2708b', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:86:b9:b3', 'vm-uuid': 'a02f7f2f-d2cf-4612-b18a-8be435257201'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.132 232622 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.133 232622 INFO os_vif [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Successfully plugged vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b')#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.133 232622 DEBUG nova.compute.manager [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.138 232622 DEBUG nova.compute.manager [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Current state is 1, state in DB is 1. _init_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:1304#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.138 232622 INFO nova.compute.manager [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.232 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.232 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.233 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.233 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.234 232622 DEBUG oslo_concurrency.processutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.688 232622 DEBUG oslo_concurrency.processutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.454s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.767 232622 DEBUG nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.767 232622 DEBUG nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.991 232622 WARNING nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.994 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=12902MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.994 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:34:36 localhost nova_compute[232618]: 2025-11-27 09:34:36.995 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.143 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.144 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.144 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.197 232622 DEBUG nova.scheduler.client.report [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.218 232622 DEBUG nova.scheduler.client.report [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.219 232622 DEBUG nova.compute.provider_tree [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.243 232622 DEBUG nova.scheduler.client.report [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.276 232622 DEBUG nova.scheduler.client.report [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: HW_CPU_X86_ABM,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_AVX2,COMPUTE_TRUSTED_CERTS,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_AMD_SVM,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_RESCUE_BFV,COMPUTE_VOLUME_ATTACH_WITH_TAG,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_AMI,HW_CPU_X86_AVX,HW_CPU_X86_BMI,COMPUTE_STORAGE_BUS_FDC,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_VIOMMU_MODEL_AUTO,COMPUTE_STORAGE_BUS_SATA,COMPUTE_NET_VIF_MODEL_E1000E,COMPUTE_IMAGE_TYPE_ARI,HW_CPU_X86_SVM,COMPUTE_NET_VIF_MODEL_VMXNET3,COMPUTE_GRAPHICS_MODEL_BOCHS,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_NET_VIF_MODEL_NE2K_PCI,HW_CPU_X86_BMI2,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_NET_VIF_MODEL_PCNET,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_SECURITY_TPM_2_0,HW_CPU_X86_FMA3,HW_CPU_X86_SSE2,HW_CPU_X86_SSE,HW_CPU_X86_AESNI,COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_GRAPHICS_MODEL_CIRRUS,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_DEVICE_TAGGING,COMPUTE_VOLUME_EXTEND,HW_CPU_X86_SSE4A,COMPUTE_STORAGE_BUS_IDE,HW_CPU_X86_MMX,HW_CPU_X86_SSE41,COMPUTE_ACCELERATORS,HW_CPU_X86_SSE42,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,HW_CPU_X86_SSSE3,HW_CPU_X86_F16C,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_VIOMMU_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_VGA,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_RTL8139,HW_CPU_X86_CLMUL,COMPUTE_STORAGE_BUS_USB _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.314 232622 DEBUG oslo_concurrency.processutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.754 232622 DEBUG oslo_concurrency.processutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.440s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.760 232622 DEBUG nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] /sys/module/kvm_amd/parameters/sev contains [N Nov 27 04:34:37 localhost nova_compute[232618]: ] _kernel_supports_amd_sev /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1803#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.761 232622 INFO nova.virt.libvirt.host [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] kernel doesn't support AMD SEV#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.762 232622 DEBUG nova.compute.provider_tree [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.763 232622 DEBUG nova.virt.libvirt.driver [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] CPU mode 'host-model' models '' was chosen, with extra flags: '' _get_guest_cpu_model_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:5396#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.784 232622 DEBUG nova.scheduler.client.report [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.809 232622 DEBUG nova.compute.resource_tracker [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.809 232622 DEBUG oslo_concurrency.lockutils [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.814s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.809 232622 DEBUG nova.service [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Creating RPC server for service compute start /usr/lib/python3.9/site-packages/nova/service.py:182#033[00m Nov 27 04:34:37 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=62132 DF PROTO=TCP SPT=51296 DPT=9101 SEQ=2673599764 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129487DB0000000001030307) Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.838 232622 DEBUG nova.service [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] Join ServiceGroup membership for this service compute start /usr/lib/python3.9/site-packages/nova/service.py:199#033[00m Nov 27 04:34:37 localhost nova_compute[232618]: 2025-11-27 09:34:37.839 232622 DEBUG nova.servicegroup.drivers.db [None req-954260e1-a275-4e40-976f-d74153c4689c - - - - - -] DB_Driver: join new ServiceGroup member np0005537446.localdomain to the compute group, service = join /usr/lib/python3.9/site-packages/nova/servicegroup/drivers/db.py:44#033[00m Nov 27 04:34:38 localhost nova_compute[232618]: 2025-11-27 09:34:38.620 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35878 DF PROTO=TCP SPT=55314 DPT=9882 SEQ=3448100400 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12948B770000000001030307) Nov 27 04:34:40 localhost nova_compute[232618]: 2025-11-27 09:34:40.291 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35880 DF PROTO=TCP SPT=55314 DPT=9882 SEQ=3448100400 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294979A0000000001030307) Nov 27 04:34:43 localhost sshd[232914]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:34:43 localhost systemd-logind[761]: New session 56 of user zuul. Nov 27 04:34:43 localhost systemd[1]: Started Session 56 of User zuul. Nov 27 04:34:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:34:43.540 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:34:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:34:43.540 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:34:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:34:43.542 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:34:43 localhost nova_compute[232618]: 2025-11-27 09:34:43.677 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:44 localhost python3.9[233025]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:34:45 localhost nova_compute[232618]: 2025-11-27 09:34:45.294 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=59282 DF PROTO=TCP SPT=56382 DPT=9102 SEQ=2291192338 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294A6090000000001030307) Nov 27 04:34:46 localhost python3.9[233139]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:34:46 localhost systemd[1]: Reloading. Nov 27 04:34:46 localhost systemd-rc-local-generator[233165]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:34:46 localhost systemd-sysv-generator[233170]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:34:46 localhost systemd[1]: tmp-crun.6AdG5i.mount: Deactivated successfully. Nov 27 04:34:46 localhost podman[233176]: 2025-11-27 09:34:46.617845582 +0000 UTC m=+0.088691394 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 04:34:46 localhost podman[233176]: 2025-11-27 09:34:46.653140376 +0000 UTC m=+0.123986208 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:34:46 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:34:47 localhost python3.9[233309]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:34:47 localhost network[233326]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:34:47 localhost network[233327]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:34:47 localhost network[233328]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:34:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12485 DF PROTO=TCP SPT=54994 DPT=9100 SEQ=1804943258 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294AE1A0000000001030307) Nov 27 04:34:48 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:34:48 localhost nova_compute[232618]: 2025-11-27 09:34:48.681 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:50 localhost nova_compute[232618]: 2025-11-27 09:34:50.299 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=12486 DF PROTO=TCP SPT=54994 DPT=9100 SEQ=1804943258 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294BDDB0000000001030307) Nov 27 04:34:52 localhost python3.9[233563]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_ceilometer_agent_compute.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:34:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:34:52 localhost podman[233565]: 2025-11-27 09:34:52.374635958 +0000 UTC m=+0.081629557 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:34:52 localhost podman[233565]: 2025-11-27 09:34:52.403920634 +0000 UTC m=+0.110914233 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:34:52 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:34:53 localhost nova_compute[232618]: 2025-11-27 09:34:53.682 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55646 DF PROTO=TCP SPT=60286 DPT=9101 SEQ=1365417040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294C5DA0000000001030307) Nov 27 04:34:53 localhost python3.9[233694]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_ceilometer_agent_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:34:53 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 76.6 (255 of 333 items), suggesting rotation. Nov 27 04:34:53 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:34:53 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:34:53 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:34:54 localhost python3.9[233805]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_ceilometer_agent_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:34:55 localhost nova_compute[232618]: 2025-11-27 09:34:55.351 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:55 localhost python3.9[233915]: ansible-ansible.legacy.command Invoked with _raw_params=if systemctl is-active certmonger.service; then#012 systemctl disable --now certmonger.service#012 test -f /etc/systemd/system/certmonger.service || systemctl mask certmonger.service#012fi#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:34:56 localhost python3.9[234025]: ansible-ansible.builtin.find Invoked with file_type=any hidden=True paths=['/var/lib/certmonger/requests'] patterns=[] read_whole_file=False age_stamp=mtime recurse=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:34:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=53754 DF PROTO=TCP SPT=53328 DPT=9101 SEQ=3153789417 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294D1DA0000000001030307) Nov 27 04:34:57 localhost python3.9[234135]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:34:57 localhost systemd[1]: Reloading. Nov 27 04:34:57 localhost systemd-rc-local-generator[234161]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:34:57 localhost systemd-sysv-generator[234165]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:57 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:34:58 localhost python3.9[234280]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_ceilometer_agent_compute.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:34:58 localhost nova_compute[232618]: 2025-11-27 09:34:58.687 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:34:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55648 DF PROTO=TCP SPT=60286 DPT=9101 SEQ=1365417040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294DD9B0000000001030307) Nov 27 04:35:00 localhost python3.9[234391]: ansible-ansible.builtin.file Invoked with group=zuul mode=0750 owner=zuul path=/var/lib/openstack/config/telemetry recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:00 localhost nova_compute[232618]: 2025-11-27 09:35:00.355 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:01 localhost python3.9[234499]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.841 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.897 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.898 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.898 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.899 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._cleanup_running_deleted_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:01 localhost nova_compute[232618]: 2025-11-27 09:35:01.955 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.056s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:35:01 localhost python3.9[234609]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer-host-specific.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:02 localhost python3.9[234695]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/ceilometer-host-specific.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236101.5242238-362-104691186699131/.source.conf follow=False _original_basename=ceilometer-host-specific.conf.j2 checksum=307bac8d43f2a95c554456dd97b1d8f5806b0a34 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:35:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=5991 DF PROTO=TCP SPT=38914 DPT=9105 SEQ=1095023020 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294E9DA0000000001030307) Nov 27 04:35:02 localhost podman[234733]: 2025-11-27 09:35:02.99667202 +0000 UTC m=+0.086778405 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:35:03 localhost podman[234733]: 2025-11-27 09:35:03.011429707 +0000 UTC m=+0.101536132 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:35:03 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:35:03 localhost python3.9[234824]: ansible-ansible.builtin.group Invoked with name=libvirt state=present force=False system=False local=False non_unique=False gid=None gid_min=None gid_max=None Nov 27 04:35:03 localhost nova_compute[232618]: 2025-11-27 09:35:03.725 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:04 localhost python3.9[234934]: ansible-ansible.builtin.getent Invoked with database=passwd key=ceilometer fail_key=True service=None split=None Nov 27 04:35:05 localhost python3.9[235045]: ansible-ansible.builtin.group Invoked with gid=42405 name=ceilometer state=present force=False system=False local=False non_unique=False gid_min=None gid_max=None Nov 27 04:35:05 localhost nova_compute[232618]: 2025-11-27 09:35:05.404 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:06 localhost python3.9[235161]: ansible-ansible.builtin.user Invoked with comment=ceilometer user group=ceilometer groups=['libvirt'] name=ceilometer shell=/sbin/nologin state=present uid=42405 non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on np0005537446.localdomain update_password=always home=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None password_expire_account_disable=None uid_min=None uid_max=None Nov 27 04:35:07 localhost python3.9[235277]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55649 DF PROTO=TCP SPT=60286 DPT=9101 SEQ=1365417040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1294FDDB0000000001030307) Nov 27 04:35:08 localhost python3.9[235363]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/ceilometer.conf mode=0640 remote_src=False src=/home/zuul/.ansible/tmp/ansible-tmp-1764236107.4726965-566-230014434497451/.source.conf _original_basename=ceilometer.conf follow=False checksum=705f6de003662ec642d109926ed2eb666d2c9797 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:08 localhost nova_compute[232618]: 2025-11-27 09:35:08.727 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22620 DF PROTO=TCP SPT=47018 DPT=9882 SEQ=4191668446 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129500A80000000001030307) Nov 27 04:35:09 localhost python3.9[235471]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/polling.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:09 localhost python3.9[235557]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/polling.yaml mode=0640 remote_src=False src=/home/zuul/.ansible/tmp/ansible-tmp-1764236108.6085794-566-39453058968216/.source.yaml _original_basename=polling.yaml follow=False checksum=6c8680a286285f2e0ef9fa528ca754765e5ed0e5 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:10 localhost python3.9[235665]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/custom.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:10 localhost nova_compute[232618]: 2025-11-27 09:35:10.453 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:10 localhost python3.9[235751]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/custom.conf mode=0640 remote_src=False src=/home/zuul/.ansible/tmp/ansible-tmp-1764236109.7247858-566-217593631224455/.source.conf _original_basename=custom.conf follow=False checksum=838b8b0a7d7f72e55ab67d39f32e3cb3eca2139b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:11 localhost python3.9[235859]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/certs/telemetry/default/tls.crt follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:35:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=22622 DF PROTO=TCP SPT=47018 DPT=9882 SEQ=4191668446 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12950C9A0000000001030307) Nov 27 04:35:12 localhost python3.9[235967]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/certs/telemetry/default/tls.key follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:35:13 localhost python3.9[236075]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:13 localhost nova_compute[232618]: 2025-11-27 09:35:13.756 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:13 localhost python3.9[236161]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236112.9098306-742-212727687335806/.source.json follow=False _original_basename=ceilometer-agent-compute.json.j2 checksum=264d11e8d3809e7ef745878dce7edd46098e25b2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:15 localhost python3.9[236269]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer-host-specific.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:15 localhost nova_compute[232618]: 2025-11-27 09:35:15.457 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:15 localhost python3.9[236324]: ansible-ansible.legacy.file Invoked with mode=420 dest=/var/lib/openstack/config/telemetry/ceilometer-host-specific.conf _original_basename=ceilometer-host-specific.conf.j2 recurse=False state=file path=/var/lib/openstack/config/telemetry/ceilometer-host-specific.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38610 DF PROTO=TCP SPT=58372 DPT=9102 SEQ=1445452414 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12951B390000000001030307) Nov 27 04:35:16 localhost python3.9[236432]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer_agent_compute.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:16 localhost python3.9[236518]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/ceilometer_agent_compute.json mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236115.6493375-742-220191557997603/.source.json follow=False _original_basename=ceilometer_agent_compute.json.j2 checksum=d15068604cf730dd6e7b88a19d62f57d3a39f94f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:35:17 localhost podman[236593]: 2025-11-27 09:35:17.001020968 +0000 UTC m=+0.086942651 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_controller, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:35:17 localhost podman[236593]: 2025-11-27 09:35:17.042796131 +0000 UTC m=+0.128717834 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:35:17 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:35:17 localhost python3.9[236640]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/ceilometer_prom_exporter.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35376 DF PROTO=TCP SPT=58546 DPT=9100 SEQ=2405785545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295235A0000000001030307) Nov 27 04:35:17 localhost python3.9[236738]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/ceilometer_prom_exporter.yaml mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236116.721557-742-253213749779501/.source.yaml follow=False _original_basename=ceilometer_prom_exporter.yaml.j2 checksum=10157c879411ee6023e506dc85a343cedc52700f backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:18 localhost python3.9[236846]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/firewall.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:18 localhost python3.9[236932]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/firewall.yaml mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236117.776083-742-162833628698006/.source.yaml follow=False _original_basename=firewall.yaml.j2 checksum=d942d984493b214bda2913f753ff68cdcedff00e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:18 localhost nova_compute[232618]: 2025-11-27 09:35:18.795 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:19 localhost python3.9[237040]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/node_exporter.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:19 localhost python3.9[237126]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/node_exporter.json mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236118.8574586-742-253379205629088/.source.json follow=False _original_basename=node_exporter.json.j2 checksum=7e5ab36b7368c1d4a00810e02af11a7f7d7c84e8 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:20 localhost python3.9[237234]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/node_exporter.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:20 localhost nova_compute[232618]: 2025-11-27 09:35:20.500 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:20 localhost python3.9[237320]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/node_exporter.yaml mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236119.919828-742-244395228119407/.source.yaml follow=False _original_basename=node_exporter.yaml.j2 checksum=81d906d3e1e8c4f8367276f5d3a67b80ca7e989e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:21 localhost python3.9[237428]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/openstack_network_exporter.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=35377 DF PROTO=TCP SPT=58546 DPT=9100 SEQ=2405785545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295331A0000000001030307) Nov 27 04:35:21 localhost python3.9[237514]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/openstack_network_exporter.json mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236121.0302851-742-265139377405325/.source.json follow=False _original_basename=openstack_network_exporter.json.j2 checksum=0e4ea521b0035bea70b7a804346a5c89364dcbc3 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:22 localhost python3.9[237622]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:35:22 localhost systemd[1]: tmp-crun.PDUE84.mount: Deactivated successfully. Nov 27 04:35:22 localhost podman[237709]: 2025-11-27 09:35:22.995749819 +0000 UTC m=+0.092345165 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:35:23 localhost podman[237709]: 2025-11-27 09:35:23.003905916 +0000 UTC m=+0.100501302 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:35:23 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:35:23 localhost python3.9[237708]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236122.0623834-742-142004448078832/.source.yaml follow=False _original_basename=openstack_network_exporter.yaml.j2 checksum=b056dcaaba7624b93826bb95ee9e82f81bde6c72 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:23 localhost python3.9[237835]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/podman_exporter.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20600 DF PROTO=TCP SPT=57242 DPT=9101 SEQ=707162064 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12953B1A0000000001030307) Nov 27 04:35:23 localhost nova_compute[232618]: 2025-11-27 09:35:23.834 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:24 localhost python3.9[237921]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/podman_exporter.json mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236123.1869204-742-54339353788992/.source.json follow=False _original_basename=podman_exporter.json.j2 checksum=885ccc6f5edd8803cb385bdda5648d0b3017b4e4 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:24 localhost python3.9[238029]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/telemetry/podman_exporter.yaml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:25 localhost python3.9[238151]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/telemetry/podman_exporter.yaml mode=420 src=/home/zuul/.ansible/tmp/ansible-tmp-1764236124.3427377-742-155970242706518/.source.yaml follow=False _original_basename=podman_exporter.yaml.j2 checksum=7ccb5eca2ff1dc337c3f3ecbbff5245af7149c47 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:25 localhost nova_compute[232618]: 2025-11-27 09:35:25.531 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9709 DF PROTO=TCP SPT=51680 DPT=9105 SEQ=882892835 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295471A0000000001030307) Nov 27 04:35:27 localhost python3.9[238310]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/healthchecks setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:28 localhost nova_compute[232618]: 2025-11-27 09:35:28.867 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:29 localhost python3.9[238420]: ansible-ansible.builtin.systemd_service Invoked with enabled=True name=podman.socket state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:35:29 localhost systemd[1]: Reloading. Nov 27 04:35:29 localhost systemd-sysv-generator[238453]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:35:29 localhost systemd-rc-local-generator[238446]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:29 localhost systemd[1]: Listening on Podman API Socket. Nov 27 04:35:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20602 DF PROTO=TCP SPT=57242 DPT=9101 SEQ=707162064 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129552DA0000000001030307) Nov 27 04:35:30 localhost python3.9[238570]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/ceilometer_agent_compute/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:30 localhost nova_compute[232618]: 2025-11-27 09:35:30.566 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:30 localhost python3.9[238658]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/ceilometer_agent_compute/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236129.8766444-1259-184843999524737/.source _original_basename=healthcheck follow=False checksum=ebb343c21fce35a02591a9351660cb7035a47d42 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:31 localhost python3.9[238713]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/ceilometer_agent_compute/healthcheck.future follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:31 localhost python3.9[238801]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/ceilometer_agent_compute/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236129.8766444-1259-184843999524737/.source.future _original_basename=healthcheck.future follow=False checksum=d500a98192f4ddd70b4dfdc059e2d81aed36a294 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=9711 DF PROTO=TCP SPT=51680 DPT=9105 SEQ=882892835 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12955EDA0000000001030307) Nov 27 04:35:32 localhost python3.9[238911]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/telemetry config_pattern=ceilometer_agent_compute.json debug=False Nov 27 04:35:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:35:33 localhost systemd[1]: tmp-crun.3LB5Ee.mount: Deactivated successfully. Nov 27 04:35:33 localhost podman[239022]: 2025-11-27 09:35:33.796459671 +0000 UTC m=+0.094978565 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:35:33 localhost podman[239022]: 2025-11-27 09:35:33.80568693 +0000 UTC m=+0.104205824 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 04:35:33 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:35:33 localhost nova_compute[232618]: 2025-11-27 09:35:33.899 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:33 localhost python3.9[239021]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:35:34 localhost nova_compute[232618]: 2025-11-27 09:35:34.331 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:34 localhost nova_compute[232618]: 2025-11-27 09:35:34.332 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:34 localhost nova_compute[232618]: 2025-11-27 09:35:34.332 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:35:34 localhost nova_compute[232618]: 2025-11-27 09:35:34.333 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:35:35 localhost python3[239149]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/telemetry config_id=edpm config_overrides={} config_patterns=ceilometer_agent_compute.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:35:35 localhost python3[239149]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "e6f07353639e492d8c9627d6d615ceeb47cb00ac4d14993b12e8023ee2aeee6f",#012 "Digest": "sha256:ba8d4a4e89620dec751cb5de5631f858557101d862972a8e817b82e4e10180a1",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:ba8d4a4e89620dec751cb5de5631f858557101d862972a8e817b82e4e10180a1"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:26:47.510377458Z",#012 "Config": {#012 "User": "root",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 505178369,#012 "VirtualSize": 505178369,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/dc5b8b4def912dce4d14a76402b323c6b5c48ee8271230eacbdaaa7e58e676b2/diff:/var/lib/containers/storage/overlay/f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/5ea32d7a444086a7f1ea2479bd7b214a5adab9651f7d4df1f24a039ae5563f9d/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/5ea32d7a444086a7f1ea2479bd7b214a5adab9651f7d4df1f24a039ae5563f9d/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:c136b33417f134a3b932677bcf7a2df089c29f20eca250129eafd2132d4708bb",#012 "sha256:df29e1f065b3ca62a976bd39a05f70336eee2ae6be8f0f1548e8c749ab2e29f2",#012 "sha256:23884b48504b714fa8c89fa23b204d39c39cc69fece546e604d8bd0566e4fb11"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "root",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 Nov 27 04:35:35 localhost nova_compute[232618]: 2025-11-27 09:35:35.608 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:35 localhost podman[239198]: 2025-11-27 09:35:35.668765511 +0000 UTC m=+0.126256270 container remove f9e69bf6cd4b4b0f084aa7b7bc5872342cd4d4c314cfd10c42d44dc0c85e0d3a (image=registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1, name=ceilometer_agent_compute, vcs-type=git, io.buildah.version=1.41.4, managed_by=tripleo_ansible, vendor=Red Hat, Inc., description=Red Hat OpenStack Platform 17.1 ceilometer-compute, build-date=2025-11-19T00:11:48Z, url=https://www.redhat.com, io.k8s.display-name=Red Hat OpenStack Platform 17.1 ceilometer-compute, baseimage=registry.redhat.io/rhel9-2-els/rhel:9.2@sha256:dd3e22348293588538689be8c51c23472fd4ca53650b3898401947ef9c7e1a05, container_name=ceilometer_agent_compute, version=17.1.12, name=rhosp17/openstack-ceilometer-compute, architecture=x86_64, org.opencontainers.image.revision=073ea4b06e5aa460399b0c251f416da40b228676, cpe=cpe:/a:redhat:rhel_e4s:9.2::appstream, config_id=tripleo_step4, distribution-scope=public, config_data={'depends_on': ['tripleo_nova_libvirt.target'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'TRIPLEO_CONFIG_HASH': 'd7f975e76c3815f85351b6258d097495'}, 'healthcheck': {'test': '/openstack/healthcheck'}, 'image': 'registry.redhat.io/rhosp-rhel9/openstack-ceilometer-compute:17.1', 'net': 'host', 'privileged': False, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/etc/puppet:/etc/puppet:ro', '/var/lib/kolla/config_files/ceilometer_agent_compute.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/config-data/puppet-generated/ceilometer:/var/lib/kolla/config_files/src:ro', '/run/libvirt:/run/libvirt:shared,z', '/var/log/containers/ceilometer:/var/log/ceilometer:z']}, tcib_managed=true, batch=17.1_20251118.1, io.openshift.tags=rhosp osp openstack osp-17.1 openstack-ceilometer-compute, vcs-ref=073ea4b06e5aa460399b0c251f416da40b228676, io.openshift.expose-services=, com.redhat.component=openstack-ceilometer-compute-container, summary=Red Hat OpenStack Platform 17.1 ceilometer-compute, maintainer=OpenStack TripleO Team, release=1761123044, io.k8s.description=Red Hat OpenStack Platform 17.1 ceilometer-compute, konflux.additional-tags=17.1.12 17.1_20251118.1, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:35:35 localhost python3[239149]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman rm --force ceilometer_agent_compute Nov 27 04:35:35 localhost podman[239212]: Nov 27 04:35:35 localhost podman[239212]: 2025-11-27 09:35:35.770600112 +0000 UTC m=+0.082758985 container create 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=edpm, managed_by=edpm_ansible, container_name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:35:35 localhost podman[239212]: 2025-11-27 09:35:35.733885062 +0000 UTC m=+0.046043985 image pull quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified Nov 27 04:35:35 localhost python3[239149]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name ceilometer_agent_compute --conmon-pidfile /run/ceilometer_agent_compute.pid --env KOLLA_CONFIG_STRATEGY=COPY_ALWAYS --env OS_ENDPOINT_TYPE=internal --healthcheck-command /openstack/healthcheck compute --label config_id=edpm --label container_name=ceilometer_agent_compute --label managed_by=edpm_ansible --label config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']} --log-driver journald --log-level info --network host --security-opt label:type:ceilometer_polling_t --user ceilometer --volume /var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z --volume /var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z --volume /run/libvirt:/run/libvirt:shared,ro --volume /etc/hosts:/etc/hosts:ro --volume /etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro --volume /etc/localtime:/etc/localtime:ro --volume /etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro --volume /var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z --volume /dev/log:/dev/log --volume /var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified kolla_start Nov 27 04:35:35 localhost nova_compute[232618]: 2025-11-27 09:35:35.836 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:35:35 localhost nova_compute[232618]: 2025-11-27 09:35:35.837 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:35:35 localhost nova_compute[232618]: 2025-11-27 09:35:35.837 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:35:35 localhost nova_compute[232618]: 2025-11-27 09:35:35.838 232622 DEBUG nova.objects.instance [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.360 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.390 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.391 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.391 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.392 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.393 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.393 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.394 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.395 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.396 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.396 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.417 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.418 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.419 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.419 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.420 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:35:36 localhost python3.9[239361]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.868 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.447s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.930 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:35:36 localhost nova_compute[232618]: 2025-11-27 09:35:36.931 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.121 232622 WARNING nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.122 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=12900MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.122 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.122 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.192 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.193 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.193 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.231 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:35:37 localhost python3.9[239495]: ansible-file Invoked with path=/etc/systemd/system/edpm_ceilometer_agent_compute.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.692 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.461s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.699 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.722 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.724 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:35:37 localhost nova_compute[232618]: 2025-11-27 09:35:37.725 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.603s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:35:38 localhost python3.9[239625]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236137.5533774-1451-275420583623606/source dest=/etc/systemd/system/edpm_ceilometer_agent_compute.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20603 DF PROTO=TCP SPT=57242 DPT=9101 SEQ=707162064 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129573DA0000000001030307) Nov 27 04:35:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16740 DF PROTO=TCP SPT=45040 DPT=9882 SEQ=3894413941 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129575D70000000001030307) Nov 27 04:35:38 localhost nova_compute[232618]: 2025-11-27 09:35:38.945 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:39 localhost python3.9[239680]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:35:39 localhost systemd[1]: Reloading. Nov 27 04:35:39 localhost systemd-rc-local-generator[239706]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:35:39 localhost systemd-sysv-generator[239709]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:39 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost python3.9[239770]: ansible-systemd Invoked with state=restarted name=edpm_ceilometer_agent_compute.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:35:40 localhost systemd[1]: Reloading. Nov 27 04:35:40 localhost systemd-rc-local-generator[239795]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:35:40 localhost systemd-sysv-generator[239800]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:40 localhost nova_compute[232618]: 2025-11-27 09:35:40.655 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:40 localhost systemd[1]: Starting ceilometer_agent_compute container... Nov 27 04:35:40 localhost systemd[1]: Started libcrun container. Nov 27 04:35:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c292b16e30cae055c2dd95c4b88b3c70bc19497e668938824a9386cf4950cb28/merged/var/lib/openstack/config supports timestamps until 2038 (0x7fffffff) Nov 27 04:35:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c292b16e30cae055c2dd95c4b88b3c70bc19497e668938824a9386cf4950cb28/merged/var/lib/kolla/config_files/config.json supports timestamps until 2038 (0x7fffffff) Nov 27 04:35:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:35:40 localhost podman[239811]: 2025-11-27 09:35:40.874936718 +0000 UTC m=+0.146981498 container init 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm) Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: + sudo -E kolla_set_configs Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: sudo: unable to send audit message: Operation not permitted Nov 27 04:35:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:35:40 localhost podman[239811]: 2025-11-27 09:35:40.918621899 +0000 UTC m=+0.190666629 container start 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 04:35:40 localhost podman[239811]: ceilometer_agent_compute Nov 27 04:35:40 localhost systemd[1]: Started ceilometer_agent_compute container. Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Validating config file Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Copying service configuration files Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Deleting /etc/ceilometer/ceilometer.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Copying /var/lib/openstack/config/ceilometer.conf to /etc/ceilometer/ceilometer.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Deleting /etc/ceilometer/polling.yaml Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Copying /var/lib/openstack/config/polling.yaml to /etc/ceilometer/polling.yaml Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Setting permission for /etc/ceilometer/polling.yaml Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Copying /var/lib/openstack/config/custom.conf to /etc/ceilometer/ceilometer.conf.d/01-ceilometer-custom.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf.d/01-ceilometer-custom.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Copying /var/lib/openstack/config/ceilometer-host-specific.conf to /etc/ceilometer/ceilometer.conf.d/02-ceilometer-host-specific.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf.d/02-ceilometer-host-specific.conf Nov 27 04:35:40 localhost ceilometer_agent_compute[239825]: INFO:__main__:Writing out command to execute Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: ++ cat /run_command Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + CMD='/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout' Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + ARGS= Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + sudo kolla_copy_cacerts Nov 27 04:35:41 localhost podman[239833]: 2025-11-27 09:35:41.017246282 +0000 UTC m=+0.093916172 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=starting, managed_by=edpm_ansible, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: sudo: unable to send audit message: Operation not permitted Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + [[ ! -n '' ]] Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + . kolla_extend_start Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + echo 'Running command: '\''/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout'\''' Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: Running command: '/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout' Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + umask 0022 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: + exec /usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout Nov 27 04:35:41 localhost podman[239833]: 2025-11-27 09:35:41.046476127 +0000 UTC m=+0.123145997 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:35:41 localhost podman[239833]: unhealthy Nov 27 04:35:41 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:35:41 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Failed with result 'exit-code'. Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.762 2 DEBUG cotyledon.oslo_config_glue [-] Full set of CONF: _load_service_manager_options /usr/lib/python3.9/site-packages/cotyledon/oslo_config_glue.py:40 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.762 2 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.762 2 DEBUG cotyledon.oslo_config_glue [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.762 2 DEBUG cotyledon.oslo_config_glue [-] command line args: ['--polling-namespaces', 'compute', '--logfile', '/dev/stdout'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.762 2 DEBUG cotyledon.oslo_config_glue [-] config files: ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] config_dir = ['/etc/ceilometer/ceilometer.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] config_file = ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'futurist=INFO', 'neutronclient=INFO', 'keystoneclient=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] event_pipeline_cfg_file = event_pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.763 2 DEBUG cotyledon.oslo_config_glue [-] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] http_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] hypervisor_inspector = libvirt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] libvirt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] libvirt_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] log_dir = /var/log/ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] log_file = /dev/stdout log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.764 2 DEBUG cotyledon.oslo_config_glue [-] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] max_parallel_requests = 64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.765 2 DEBUG cotyledon.oslo_config_glue [-] partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] pipeline_cfg_file = pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] polling_namespaces = ['compute'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] reseller_prefix = AUTH_ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_keys = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_length = 256 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.766 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_namespace = ['metering.'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] rootwrap_config = /etc/ceilometer/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] sample_source = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.767 2 DEBUG cotyledon.oslo_config_glue [-] compute.instance_discovery_method = libvirt_metadata log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] compute.resource_cache_expiry = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] compute.resource_update_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] coordination.backend_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] event.definitions_cfg_file = event_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] event.drop_unmatched_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] event.store_raw = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] ipmi.node_manager_init_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] ipmi.polling_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] meter.meter_definitions_dirs = ['/etc/ceilometer/meters.d', '/usr/lib/python3.9/site-packages/ceilometer/data/meters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_path = mon_pub_failures.txt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.768 2 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_count = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_polling_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_timeout = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.client_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.769 2 DEBUG cotyledon.oslo_config_glue [-] monasca.client_retry_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.clientapi_version = 2_0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cloud_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cluster = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.control_plane = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.enable_api_pagination = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.770 2 DEBUG cotyledon.oslo_config_glue [-] monasca.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.monasca_mappings = /etc/ceilometer/monasca_field_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.retry_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] monasca.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] notification.ack_on_event_error = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.771 2 DEBUG cotyledon.oslo_config_glue [-] notification.batch_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] notification.batch_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] notification.messaging_urls = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] notification.notification_control_exchanges = ['nova', 'glance', 'neutron', 'cinder', 'heat', 'keystone', 'sahara', 'trove', 'zaqar', 'swift', 'ceilometer', 'magnum', 'dns', 'ironic', 'aodh'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] notification.pipelines = ['meter', 'event'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] notification.workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] polling.batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.772 2 DEBUG cotyledon.oslo_config_glue [-] polling.cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] polling.partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] polling.pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] polling.tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] publisher.telemetry_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.event_topic = event log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.metering_topic = metering log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.telemetry_driver = messagingv2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.773 2 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.access_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.secret_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] rgw_client.implicit_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.cinder = volumev3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.glance = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.neutron = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.nova = compute log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.radosgw = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.774 2 DEBUG cotyledon.oslo_config_glue [-] service_types.swift = object-store log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_ip = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.775 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] vmware.wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.776 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.interface = internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.777 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.778 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.779 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.780 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.780 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.780 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.780 2 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613 Nov 27 04:35:41 localhost systemd[1]: tmp-crun.JEwFv7.mount: Deactivated successfully. Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.796 12 INFO ceilometer.polling.manager [-] Looking for dynamic pollsters configurations at [['/etc/ceilometer/pollsters.d']]. Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.798 12 INFO ceilometer.polling.manager [-] No dynamic pollsters found in folder [/etc/ceilometer/pollsters.d]. Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.799 12 INFO ceilometer.polling.manager [-] No dynamic pollsters file found in dirs [['/etc/ceilometer/pollsters.d']]. Nov 27 04:35:41 localhost python3.9[239965]: ansible-ansible.builtin.systemd Invoked with name=edpm_ceilometer_agent_compute.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:35:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=16742 DF PROTO=TCP SPT=45040 DPT=9882 SEQ=3894413941 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129581DA0000000001030307) Nov 27 04:35:41 localhost systemd[1]: Stopping ceilometer_agent_compute container... Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.895 12 DEBUG ceilometer.compute.virt.libvirt.utils [-] Connecting to libvirt: qemu:///system new_libvirt_connection /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/utils.py:93 Nov 27 04:35:41 localhost systemd[1]: tmp-crun.7jXgn5.mount: Deactivated successfully. Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.954 2 INFO cotyledon._service_manager [-] Caught SIGTERM signal, graceful exiting of master process Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] Full set of CONF: _load_service_options /usr/lib/python3.9/site-packages/cotyledon/oslo_config_glue.py:48 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] command line args: ['--polling-namespaces', 'compute', '--logfile', '/dev/stdout'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] config files: ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.958 12 DEBUG cotyledon.oslo_config_glue [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] config_dir = ['/etc/ceilometer/ceilometer.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] config_file = ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] control_exchange = ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'futurist=INFO', 'neutronclient=INFO', 'keystoneclient=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] event_pipeline_cfg_file = event_pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.959 12 DEBUG cotyledon.oslo_config_glue [-] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] http_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] hypervisor_inspector = libvirt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] libvirt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] libvirt_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] log_dir = /var/log/ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] log_file = /dev/stdout log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.960 12 DEBUG cotyledon.oslo_config_glue [-] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] max_parallel_requests = 64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.961 12 DEBUG cotyledon.oslo_config_glue [-] partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] pipeline_cfg_file = pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] polling_namespaces = ['compute'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] reseller_prefix = AUTH_ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_keys = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_length = 256 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_namespace = ['metering.'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.962 12 DEBUG cotyledon.oslo_config_glue [-] rootwrap_config = /etc/ceilometer/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] sample_source = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.963 12 DEBUG cotyledon.oslo_config_glue [-] compute.instance_discovery_method = libvirt_metadata log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] compute.resource_cache_expiry = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] compute.resource_update_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] coordination.backend_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] event.definitions_cfg_file = event_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] event.drop_unmatched_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] event.store_raw = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] ipmi.node_manager_init_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] ipmi.polling_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] meter.meter_definitions_dirs = ['/etc/ceilometer/meters.d', '/usr/lib/python3.9/site-packages/ceilometer/data/meters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_path = mon_pub_failures.txt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.964 12 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_count = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_polling_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_timeout = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.client_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.client_retry_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.965 12 DEBUG cotyledon.oslo_config_glue [-] monasca.clientapi_version = 2_0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cloud_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cluster = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.control_plane = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.enable_api_pagination = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.monasca_mappings = /etc/ceilometer/monasca_field_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.retry_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.966 12 DEBUG cotyledon.oslo_config_glue [-] monasca.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] monasca.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.ack_on_event_error = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.batch_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.batch_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.messaging_urls = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.notification_control_exchanges = ['nova', 'glance', 'neutron', 'cinder', 'heat', 'keystone', 'sahara', 'trove', 'zaqar', 'swift', 'ceilometer', 'magnum', 'dns', 'ironic', 'aodh'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.pipelines = ['meter', 'event'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] notification.workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] polling.batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] polling.cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.967 12 DEBUG cotyledon.oslo_config_glue [-] polling.partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] polling.pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] polling.tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] publisher.telemetry_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.event_topic = event log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.metering_topic = metering log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.telemetry_driver = messagingv2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.access_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.secret_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] rgw_client.implicit_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] service_types.cinder = volumev3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.968 12 DEBUG cotyledon.oslo_config_glue [-] service_types.glance = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] service_types.neutron = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] service_types.nova = compute log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] service_types.radosgw = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] service_types.swift = object-store log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_ip = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.969 12 DEBUG cotyledon.oslo_config_glue [-] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] vmware.wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.970 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.interface = internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_name = service log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.system_scope = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.971 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.username = ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.972 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.973 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.driver = ['noop'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.amqp_durable_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.974 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.975 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_queue = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.976 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.977 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.977 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.977 12 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.977 12 DEBUG cotyledon._service [-] Run service AgentManager(0) [12] wait_forever /usr/lib/python3.9/site-packages/cotyledon/_service.py:241 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.978 12 DEBUG ceilometer.agent [-] Config file: {'sources': [{'name': 'pollsters', 'interval': 120, 'meters': ['power.state', 'cpu', 'memory.usage', 'disk.*', 'network.*']}]} load_config /usr/lib/python3.9/site-packages/ceilometer/agent.py:64 Nov 27 04:35:41 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:41.983 12 DEBUG ceilometer.compute.virt.libvirt.utils [-] Connecting to libvirt: qemu:///system new_libvirt_connection /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/utils.py:93 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.055 2 DEBUG cotyledon._service_manager [-] Killing services with signal SIGTERM _shutdown /usr/lib/python3.9/site-packages/cotyledon/_service_manager.py:304 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.056 2 DEBUG cotyledon._service_manager [-] Waiting services to terminate _shutdown /usr/lib/python3.9/site-packages/cotyledon/_service_manager.py:308 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.056 12 INFO cotyledon._service [-] Caught SIGTERM signal, graceful exiting of service AgentManager(0) [12] Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.344 12 DEBUG novaclient.v2.client [-] REQ: curl -g -i -X GET http://nova-internal.openstack.svc:8774/v2.1/flavors?is_public=None -H "Accept: application/json" -H "User-Agent: python-novaclient" -H "X-Auth-Token: {SHA256}f8cdfdd9b4e33661625a21932ce1ce9fa1d13db543c5620685491c15258a1b6d" -H "X-OpenStack-Nova-API-Version: 2.1" _http_log_request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:519 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.460 12 DEBUG novaclient.v2.client [-] RESP: [200] Connection: Keep-Alive Content-Length: 327 Content-Type: application/json Date: Thu, 27 Nov 2025 09:35:42 GMT Keep-Alive: timeout=5, max=100 OpenStack-API-Version: compute 2.1 Server: Apache Vary: OpenStack-API-Version,X-OpenStack-Nova-API-Version X-OpenStack-Nova-API-Version: 2.1 x-compute-request-id: req-579bfa3f-9ed6-4cc3-9792-ea4270c9783d x-openstack-request-id: req-579bfa3f-9ed6-4cc3-9792-ea4270c9783d _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:550 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.460 12 DEBUG novaclient.v2.client [-] RESP BODY: {"flavors": [{"id": "ad8d4a49-18f7-4954-9703-7480b3ae8896", "name": "m1.small", "links": [{"rel": "self", "href": "http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}, {"rel": "bookmark", "href": "http://nova-internal.openstack.svc:8774/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}]}]} _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:582 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.460 12 DEBUG novaclient.v2.client [-] GET call to compute for http://nova-internal.openstack.svc:8774/v2.1/flavors?is_public=None used request id req-579bfa3f-9ed6-4cc3-9792-ea4270c9783d request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:954 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.462 12 DEBUG novaclient.v2.client [-] REQ: curl -g -i -X GET http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896 -H "Accept: application/json" -H "User-Agent: python-novaclient" -H "X-Auth-Token: {SHA256}f8cdfdd9b4e33661625a21932ce1ce9fa1d13db543c5620685491c15258a1b6d" -H "X-OpenStack-Nova-API-Version: 2.1" _http_log_request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:519 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.505 12 DEBUG novaclient.v2.client [-] RESP: [200] Connection: Keep-Alive Content-Length: 494 Content-Type: application/json Date: Thu, 27 Nov 2025 09:35:42 GMT Keep-Alive: timeout=5, max=99 OpenStack-API-Version: compute 2.1 Server: Apache Vary: OpenStack-API-Version,X-OpenStack-Nova-API-Version X-OpenStack-Nova-API-Version: 2.1 x-compute-request-id: req-7a965dad-7fd5-42b1-9c4c-380d3c6b1e9e x-openstack-request-id: req-7a965dad-7fd5-42b1-9c4c-380d3c6b1e9e _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:550 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.505 12 DEBUG novaclient.v2.client [-] RESP BODY: {"flavor": {"id": "ad8d4a49-18f7-4954-9703-7480b3ae8896", "name": "m1.small", "ram": 512, "disk": 1, "swap": "", "OS-FLV-EXT-DATA:ephemeral": 1, "OS-FLV-DISABLED:disabled": false, "vcpus": 1, "os-flavor-access:is_public": true, "rxtx_factor": 1.0, "links": [{"rel": "self", "href": "http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}, {"rel": "bookmark", "href": "http://nova-internal.openstack.svc:8774/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}]}} _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:582 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.505 12 DEBUG novaclient.v2.client [-] GET call to compute for http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896 used request id req-7a965dad-7fd5-42b1-9c4c-380d3c6b1e9e request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:954 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.507 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.507 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.512 12 DEBUG ceilometer.compute.virt.libvirt.inspector [-] No delta meter predecessor for a02f7f2f-d2cf-4612-b18a-8be435257201 / tapa2718872-3b inspect_vnics /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/inspector.py:136 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.513 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 11314 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '79be9468-8fe8-4ece-958a-104cf117bafe', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 11314, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.508035', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '713fd9ca-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '9bd6e07a26bf4afef82c9ad2f6a2204069d8dd5f6c7c29cebf7fc2eac4108210'}]}, 'timestamp': '2025-11-27 09:35:42.514409', '_unique_id': '75e634adc44a4e56a377bd8b8fabf0d8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.521 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.525 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.565 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 203265086 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.566 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 25604091 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c8550d02-09ab-4a44-a60a-68fd132e97a9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 203265086, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.525397', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '7147c02c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '15396a37e3947f4d604eb00690ffca55125f156773d28a2da9c1b0042ade96a3'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 25604091, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.525397', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '7147d2ba-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '8c13e0139df359921f815419ba22c5b6d7bd9ab6d92374a4b2c65b594bcd9941'}]}, 'timestamp': '2025-11-27 09:35:42.566482', '_unique_id': '6a3332d257854f9ab1fbfc5a67719d2e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.567 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.568 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.569 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 8825 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '563a1ca1-e173-441d-a370-75f4a3e7fd9b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 8825, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.569128', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '71484bb4-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '779705deaa7ebfebb092fc1500d85fd2ba4c125e1925a65f2c73923f07fb763d'}]}, 'timestamp': '2025-11-27 09:35:42.569635', '_unique_id': '0f7e2dc81f594c36a2a10343e269d4e7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.570 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.571 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.rate in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.571 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for IncomingBytesRatePollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.572 12 ERROR ceilometer.polling.manager [-] Prevent pollster network.incoming.bytes.rate from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.572 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.572 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 130 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2ce3c7ea-cf9f-4cb7-85b0-d687fd327bcc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 130, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.572705', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '7148d778-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '1b9a13feefba7e8551945f4bd1f64236b77824bc37f0005482be0c281aff9595'}]}, 'timestamp': '2025-11-27 09:35:42.573180', '_unique_id': '8d4e093a04d344dab830d05a0cb7c240'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.574 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.575 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.575 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '40b061eb-0856-4759-b54e-023410e97caa', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.575335', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '71493f6a-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '6991d0eed5f099b852586fc8ead3a3a82193948a6393a1a259b3e7081233a40c'}]}, 'timestamp': '2025-11-27 09:35:42.575866', '_unique_id': '81fa87e0a6f94fe99930957e8b4c80ef'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.576 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.577 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.578 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9b330f58-7cb4-4fd3-94c6-fcc6ce3d9371', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.577968', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '7149a48c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '56eea8a3b65d2577bd27c8b1d2da7d57c93744ecb80b43125baba17436ffa096'}]}, 'timestamp': '2025-11-27 09:35:42.578426', '_unique_id': '6bc04f6a2f27423184023bfd4fda9380'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.579 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.580 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.580 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 29130240 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.581 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 4300800 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8985ef9e-8358-4ceb-ac7b-e8402f19c000', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 29130240, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.580652', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '714a0d3c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '1e5d3fedc29f89be9f40cdf2c1bdbb9aa758e8a2c4068fe0eac8ca3d05c26dac'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 4300800, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.580652', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '714a1dae-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': 'd1ebb5ce572b04ec160b37be1fcc285dcba1c2c0630b075e287407e557e4e34c'}]}, 'timestamp': '2025-11-27 09:35:42.581501', '_unique_id': '9b953e7ef78d4d5aaa6c9a4db27832a7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.582 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.583 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.583 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1064 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.584 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 222 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f746d721-efb5-4849-b9d4-cea318ba424a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1064, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.583780', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '714a8924-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': 'a1d84a68d8ec117dd25b8a62647d595e2249179022b3223f2c430850b258489f'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 222, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.583780', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '714a996e-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '9b7c451f7a0e0915b260323c467042ddf8532d71dfb40f8fc5d5fcf262bfcef6'}]}, 'timestamp': '2025-11-27 09:35:42.584699', '_unique_id': '40494e1dae634d3aa2f52f9ea3600b4a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.585 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.586 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.587 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f0c2bc2f-d9a4-48a2-ae60-7c741e1b34bd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.587015', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '714b064c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': 'edc082ed14f0b466788259f163ebe545abc17b631ed60a0127957a714b7eeb22'}]}, 'timestamp': '2025-11-27 09:35:42.587483', '_unique_id': '726d58586047469f8f7a8998405d77b1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.588 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.589 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.589 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 974916304 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.590 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 177343418 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '36bae5a2-e42f-4f99-b735-df96c8202cb0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 974916304, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.589640', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '714b6c4a-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': 'f180f6e4a929b9ce07778728bf7d19465b8cd7b2d722e45a069b0e6948ae49af'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 177343418, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.589640', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '714b7c58-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': 'c1f7cd45339214353a3902a4d6945426ac5e55775d26caf9af6abb5b08f5c0c6'}]}, 'timestamp': '2025-11-27 09:35:42.590474', '_unique_id': '176a0d47a5bf4ef685119e0da5d71e48'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.591 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.592 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.593 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e40a1b5f-d53b-4d76-ad91-ee73d86dbaac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.592968', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '714beeae-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '819e6a1871b5483843ddd6565644f523b24783d3bebf24fb33631382531258d3'}]}, 'timestamp': '2025-11-27 09:35:42.593424', '_unique_id': '0e0b3608ea624b5287d54c4a7adddcda'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.594 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.595 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.rate in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.595 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for OutgoingBytesRatePollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.595 12 ERROR ceilometer.polling.manager [-] Prevent pollster network.outgoing.bytes.rate from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.596 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.latency in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.596 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for PerDeviceDiskLatencyPollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.596 12 ERROR ceilometer.polling.manager [-] Prevent pollster disk.device.latency from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.596 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.596 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e3c23ad0-f57a-4f62-9993-ed8e7d13bbde', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.596553', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '714c778e-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': 'be6cd647a117d3d8d92b917bf88e7af646164ef2213b2ca36674e8ba981e4329'}]}, 'timestamp': '2025-11-27 09:35:42.596843', '_unique_id': '58a9b831d7b9403582bf2c8fdd2a9ee7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.597 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.598 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.608 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.608 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c0aeb2c6-d0c8-4c6a-a029-2f4ec53eeb47', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.598120', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '714e4848-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': '69c49071dd32ba35e3e4f79f74e34e9adc329a5530bd04f6c86d33c4f6736715'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.598120', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '714e5298-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': '753da9a78f5d056dee529f2f8137c88d622e6589c14d1936fd43724d745ea190'}]}, 'timestamp': '2025-11-27 09:35:42.608985', '_unique_id': 'cb3b45f7d89a4a90b11867c728659ff0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.iops in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for PerDeviceDiskIOPSPollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 ERROR ceilometer.polling.manager [-] Prevent pollster disk.device.iops from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.610 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '03d14422-3bd9-4d07-a93f-1c87def37ad2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.610675', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '714e9f0a-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': '926a5fd1472bcf99ec7efc8fe61a6203edc46f1ecb8fb7b3657cd8a3f7fd10d9'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.610675', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '714ea900-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': '0cb87ddd5352d5a0dfa6f328dc458e5c42e8bac9a9293e7bd68d344ad7f8cc34'}]}, 'timestamp': '2025-11-27 09:35:42.611201', '_unique_id': 'fade6fce49894bdf8cadcea2d4d053ab'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.611 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.612 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.612 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'edb7e5d6-83f6-42c0-b03a-a924607e23a0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.612557', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '714ee88e-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '8d2f85f61352f4d694e2be105d4f495f78375d4459ede29bacdca8e059f84ead'}]}, 'timestamp': '2025-11-27 09:35:42.612840', '_unique_id': 'a50ef6b78b434fcd9515ae8b4b831441'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.613 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.614 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.630 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 52.46875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'dbcbbb7b-e07f-4047-b418-ca8a19c2c8fa', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 52.46875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:35:42.614199', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '7151a27c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.818172909, 'message_signature': 'd5617f37bb60c7c365de384f7e0d01ff1a0fdb65acb6ec57faecfe85ffa92c62'}]}, 'timestamp': '2025-11-27 09:35:42.630704', '_unique_id': 'f0eb623200434cb4b2c2951e830a028c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.631 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.632 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 498 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.632 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a4bf8d4b-0732-49cb-81e3-16e217f49d15', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 498, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.632064', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '7151e264-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '58d28fdd9b34ca4b42e597e8da7dbdebb4dae5989129361018a54bc1086e5406'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.632064', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '7151ebec-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '68b9fac59212fbb3261b37c474677b4d0aa570a4539ff01a1199b4d0b78abc81'}]}, 'timestamp': '2025-11-27 09:35:42.632593', '_unique_id': '8ce42df67c0341c8896a2bc19b591831'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.633 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '27a9d3a4-91e7-4285-8de8-5cb80a83d980', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.633912', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '71522a6c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': 'd5f9094814b98fa45061f5404eaee5e269b8d282bde74101a86f8a00a7d391b5'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.633912', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '71523444-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.786258634, 'message_signature': '333ea94b4a6e575371387eaabadc06aa8d2d9bebe85bc5701f835494c643ed31'}]}, 'timestamp': '2025-11-27 09:35:42.634424', '_unique_id': '20d96f86ad3c48d684698419556f3af7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.634 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.635 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.635 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 83 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2630c5f2-4db8-4fd0-a7bb-b450a7f98c9f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 83, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:42.635748', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '71527364-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.696203199, 'message_signature': '52789ebb46b5ad5dcff1ae393a23ac8959df5e728ce50f29c698d2554489a97a'}]}, 'timestamp': '2025-11-27 09:35:42.636057', '_unique_id': 'ffc96c2b7ded47e5b8728b373f95cace'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.636 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.637 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.637 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 48270000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f268dbb5-8332-459e-85d5-57ab2436394f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 48270000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:35:42.637334', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '7152b018-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.818172909, 'message_signature': '3039e34ee0b4d032885eca8585753463e060bfc48cbda80dd3e07cd94f610883'}]}, 'timestamp': '2025-11-27 09:35:42.637622', '_unique_id': '3daa8e07300e411b93dc7e5e77230a02'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.638 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 73908224 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '845a5fe1-902c-4182-92dc-af570ef645d5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 73908224, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:42.638884', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '7152eca4-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '98376902edce0c8528a8ee25ed811649103eda6c0222a975ad5cb42f60573c75'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:42.638884', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '7152f67c-cb74-11f0-b6a9-fa163e6bfee1', 'monotonic_time': 10372.713645497, 'message_signature': '81d841d15896570e0a1787540f3ad7158d6f71cef55fd83136d7d4067588032d'}]}, 'timestamp': '2025-11-27 09:35:42.639394', '_unique_id': '754cc8c1fd0647babe0b1d35b5102bee'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.639 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:42 localhost ceilometer_agent_compute[239825]: 2025-11-27 09:35:42.647 2 DEBUG cotyledon._service_manager [-] Shutdown finish _shutdown /usr/lib/python3.9/site-packages/cotyledon/_service_manager.py:320 Nov 27 04:35:42 localhost journal[205316]: End of file while reading data: Input/output error Nov 27 04:35:42 localhost journal[205316]: End of file while reading data: Input/output error Nov 27 04:35:42 localhost systemd[1]: libpod-9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.scope: Deactivated successfully. Nov 27 04:35:42 localhost systemd[1]: libpod-9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.scope: Consumed 1.315s CPU time. Nov 27 04:35:42 localhost podman[239972]: 2025-11-27 09:35:42.784935959 +0000 UTC m=+0.900886965 container died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, tcib_managed=true, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:35:42 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.timer: Deactivated successfully. Nov 27 04:35:42 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:35:42 localhost systemd[1]: tmp-crun.kkvVcH.mount: Deactivated successfully. Nov 27 04:35:42 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db-userdata-shm.mount: Deactivated successfully. Nov 27 04:35:42 localhost systemd[1]: var-lib-containers-storage-overlay-c292b16e30cae055c2dd95c4b88b3c70bc19497e668938824a9386cf4950cb28-merged.mount: Deactivated successfully. Nov 27 04:35:42 localhost podman[239972]: 2025-11-27 09:35:42.843589032 +0000 UTC m=+0.959540028 container cleanup 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=edpm, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:35:42 localhost podman[239972]: ceilometer_agent_compute Nov 27 04:35:42 localhost podman[240003]: 2025-11-27 09:35:42.938856274 +0000 UTC m=+0.064880853 container cleanup 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 04:35:42 localhost podman[240003]: ceilometer_agent_compute Nov 27 04:35:42 localhost systemd[1]: edpm_ceilometer_agent_compute.service: Deactivated successfully. Nov 27 04:35:42 localhost systemd[1]: Stopped ceilometer_agent_compute container. Nov 27 04:35:42 localhost systemd[1]: Starting ceilometer_agent_compute container... Nov 27 04:35:43 localhost systemd[1]: Started libcrun container. Nov 27 04:35:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c292b16e30cae055c2dd95c4b88b3c70bc19497e668938824a9386cf4950cb28/merged/var/lib/openstack/config supports timestamps until 2038 (0x7fffffff) Nov 27 04:35:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c292b16e30cae055c2dd95c4b88b3c70bc19497e668938824a9386cf4950cb28/merged/var/lib/kolla/config_files/config.json supports timestamps until 2038 (0x7fffffff) Nov 27 04:35:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:35:43 localhost podman[240014]: 2025-11-27 09:35:43.097860225 +0000 UTC m=+0.127192349 container init 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + sudo -E kolla_set_configs Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: sudo: unable to send audit message: Operation not permitted Nov 27 04:35:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:35:43 localhost podman[240014]: 2025-11-27 09:35:43.130013478 +0000 UTC m=+0.159345562 container start 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:35:43 localhost podman[240014]: ceilometer_agent_compute Nov 27 04:35:43 localhost systemd[1]: Started ceilometer_agent_compute container. Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Validating config file Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Copying service configuration files Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Deleting /etc/ceilometer/ceilometer.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Copying /var/lib/openstack/config/ceilometer.conf to /etc/ceilometer/ceilometer.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Deleting /etc/ceilometer/polling.yaml Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Copying /var/lib/openstack/config/polling.yaml to /etc/ceilometer/polling.yaml Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Setting permission for /etc/ceilometer/polling.yaml Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Deleting /etc/ceilometer/ceilometer.conf.d/01-ceilometer-custom.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Copying /var/lib/openstack/config/custom.conf to /etc/ceilometer/ceilometer.conf.d/01-ceilometer-custom.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf.d/01-ceilometer-custom.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Deleting /etc/ceilometer/ceilometer.conf.d/02-ceilometer-host-specific.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Copying /var/lib/openstack/config/ceilometer-host-specific.conf to /etc/ceilometer/ceilometer.conf.d/02-ceilometer-host-specific.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Setting permission for /etc/ceilometer/ceilometer.conf.d/02-ceilometer-host-specific.conf Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: INFO:__main__:Writing out command to execute Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: ++ cat /run_command Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + CMD='/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout' Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + ARGS= Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + sudo kolla_copy_cacerts Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: sudo: unable to send audit message: Operation not permitted Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + [[ ! -n '' ]] Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + . kolla_extend_start Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + echo 'Running command: '\''/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout'\''' Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: Running command: '/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout' Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + umask 0022 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: + exec /usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout Nov 27 04:35:43 localhost podman[240037]: 2025-11-27 09:35:43.221640979 +0000 UTC m=+0.085553918 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=starting, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_id=edpm, managed_by=edpm_ansible) Nov 27 04:35:43 localhost podman[240037]: 2025-11-27 09:35:43.250501073 +0000 UTC m=+0.114414082 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=edpm, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:35:43 localhost podman[240037]: unhealthy Nov 27 04:35:43 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:35:43 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Failed with result 'exit-code'. Nov 27 04:35:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:35:43.541 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:35:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:35:43.542 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:35:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:35:43.544 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:35:43 localhost python3.9[240166]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/node_exporter/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] Full set of CONF: _load_service_manager_options /usr/lib/python3.9/site-packages/cotyledon/oslo_config_glue.py:40 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] command line args: ['--polling-namespaces', 'compute', '--logfile', '/dev/stdout'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] config files: ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] config_dir = ['/etc/ceilometer/ceilometer.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.910 2 DEBUG cotyledon.oslo_config_glue [-] config_file = ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'futurist=INFO', 'neutronclient=INFO', 'keystoneclient=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] event_pipeline_cfg_file = event_pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] http_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] hypervisor_inspector = libvirt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] libvirt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.911 2 DEBUG cotyledon.oslo_config_glue [-] libvirt_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_dir = /var/log/ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_file = /dev/stdout log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.912 2 DEBUG cotyledon.oslo_config_glue [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] max_parallel_requests = 64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] pipeline_cfg_file = pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] polling_namespaces = ['compute'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] reseller_prefix = AUTH_ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.913 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_keys = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_length = 256 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_namespace = ['metering.'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] rootwrap_config = /etc/ceilometer/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] sample_source = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.914 2 DEBUG cotyledon.oslo_config_glue [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] compute.instance_discovery_method = libvirt_metadata log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] compute.resource_cache_expiry = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] compute.resource_update_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] coordination.backend_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] event.definitions_cfg_file = event_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] event.drop_unmatched_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] event.store_raw = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] ipmi.node_manager_init_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] ipmi.polling_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] meter.meter_definitions_dirs = ['/etc/ceilometer/meters.d', '/usr/lib/python3.9/site-packages/ceilometer/data/meters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.915 2 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_path = mon_pub_failures.txt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_count = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_polling_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_timeout = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.client_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.client_retry_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.916 2 DEBUG cotyledon.oslo_config_glue [-] monasca.clientapi_version = 2_0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cloud_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.cluster = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.control_plane = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.enable_api_pagination = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.monasca_mappings = /etc/ceilometer/monasca_field_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.retry_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.917 2 DEBUG cotyledon.oslo_config_glue [-] monasca.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.ack_on_event_error = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.batch_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.batch_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.messaging_urls = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.notification_control_exchanges = ['nova', 'glance', 'neutron', 'cinder', 'heat', 'keystone', 'sahara', 'trove', 'zaqar', 'swift', 'ceilometer', 'magnum', 'dns', 'ironic', 'aodh'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.pipelines = ['meter', 'event'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] notification.workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] polling.batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] polling.cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] polling.partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] polling.pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.918 2 DEBUG cotyledon.oslo_config_glue [-] polling.tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] publisher.telemetry_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.event_topic = event log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.metering_topic = metering log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.telemetry_driver = messagingv2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.access_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.secret_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] rgw_client.implicit_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] service_types.cinder = volumev3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] service_types.glance = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] service_types.neutron = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] service_types.nova = compute log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.919 2 DEBUG cotyledon.oslo_config_glue [-] service_types.radosgw = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] service_types.swift = object-store log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_ip = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.host_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] vmware.wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.920 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.interface = internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] service_credentials.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.921 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] gnocchi.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.922 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] zaqar.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.923 2 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613 Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.938 12 INFO ceilometer.polling.manager [-] Looking for dynamic pollsters configurations at [['/etc/ceilometer/pollsters.d']]. Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.939 12 INFO ceilometer.polling.manager [-] No dynamic pollsters found in folder [/etc/ceilometer/pollsters.d]. Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.940 12 INFO ceilometer.polling.manager [-] No dynamic pollsters file found in dirs [['/etc/ceilometer/pollsters.d']]. Nov 27 04:35:43 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:43.951 12 DEBUG ceilometer.compute.virt.libvirt.utils [-] Connecting to libvirt: qemu:///system new_libvirt_connection /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/utils.py:93 Nov 27 04:35:43 localhost nova_compute[232618]: 2025-11-27 09:35:43.995 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.090 12 DEBUG cotyledon.oslo_config_glue [-] Full set of CONF: _load_service_options /usr/lib/python3.9/site-packages/cotyledon/oslo_config_glue.py:48 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.090 12 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.090 12 DEBUG cotyledon.oslo_config_glue [-] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.090 12 DEBUG cotyledon.oslo_config_glue [-] command line args: ['--polling-namespaces', 'compute', '--logfile', '/dev/stdout'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.090 12 DEBUG cotyledon.oslo_config_glue [-] config files: ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] config_dir = ['/etc/ceilometer/ceilometer.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] config_file = ['/etc/ceilometer/ceilometer.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] control_exchange = ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'futurist=INFO', 'neutronclient=INFO', 'keystoneclient=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] event_pipeline_cfg_file = event_pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.091 12 DEBUG cotyledon.oslo_config_glue [-] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] http_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] hypervisor_inspector = libvirt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] libvirt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] libvirt_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] log_dir = /var/log/ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] log_file = /dev/stdout log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.092 12 DEBUG cotyledon.oslo_config_glue [-] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] log_rotation_type = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] max_logfile_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] max_logfile_size_mb = 200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] max_parallel_requests = 64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.093 12 DEBUG cotyledon.oslo_config_glue [-] pipeline_cfg_file = pipeline.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] polling_namespaces = ['compute'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] reseller_prefix = AUTH_ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_keys = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_length = 256 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] reserved_metadata_namespace = ['metering.'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] rootwrap_config = /etc/ceilometer/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.094 12 DEBUG cotyledon.oslo_config_glue [-] sample_source = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] compute.instance_discovery_method = libvirt_metadata log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] compute.resource_cache_expiry = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.095 12 DEBUG cotyledon.oslo_config_glue [-] compute.resource_update_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] coordination.backend_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] event.definitions_cfg_file = event_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] event.drop_unmatched_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] event.store_raw = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] ipmi.node_manager_init_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] ipmi.polling_retry = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] meter.meter_definitions_dirs = ['/etc/ceilometer/meters.d', '/usr/lib/python3.9/site-packages/ceilometer/data/meters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] monasca.archive_path = mon_pub_failures.txt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.096 12 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_count = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_mode = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_polling_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.batch_timeout = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.client_max_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.client_retry_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.clientapi_version = 2_0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cloud_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.097 12 DEBUG cotyledon.oslo_config_glue [-] monasca.cluster = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.control_plane = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.enable_api_pagination = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.monasca_mappings = /etc/ceilometer/monasca_field_definitions.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.retry_on_failure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] monasca.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.098 12 DEBUG cotyledon.oslo_config_glue [-] notification.ack_on_event_error = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.batch_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.batch_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.messaging_urls = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.notification_control_exchanges = ['nova', 'glance', 'neutron', 'cinder', 'heat', 'keystone', 'sahara', 'trove', 'zaqar', 'swift', 'ceilometer', 'magnum', 'dns', 'ironic', 'aodh'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.pipelines = ['meter', 'event'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] notification.workers = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] polling.batch_size = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] polling.cfg_file = polling.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] polling.partitioning_group_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] polling.pollsters_definitions_dirs = ['/etc/ceilometer/pollsters.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] polling.tenant_name_discovery = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.099 12 DEBUG cotyledon.oslo_config_glue [-] publisher.telemetry_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.event_topic = event log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.metering_topic = metering log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] publisher_notifier.telemetry_driver = messagingv2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.access_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] rgw_admin_credentials.secret_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] rgw_client.implicit_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] service_types.cinder = volumev3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] service_types.glance = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] service_types.neutron = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] service_types.nova = compute log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.100 12 DEBUG cotyledon.oslo_config_glue [-] service_types.radosgw = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] service_types.swift = object-store log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_ip = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.host_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] vmware.wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.101 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.interface = internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.102 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.project_name = service log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.system_scope = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.103 12 DEBUG cotyledon.oslo_config_glue [-] service_credentials.username = ceilometer log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] gnocchi.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_section = service_credentials log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.104 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.interface = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] zaqar.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.driver = ['noop'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.105 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.amqp_durable_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.106 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_quorum_queue = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.107 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon.oslo_config_glue [-] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.108 12 DEBUG cotyledon._service [-] Run service AgentManager(0) [12] wait_forever /usr/lib/python3.9/site-packages/cotyledon/_service.py:241 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.112 12 DEBUG ceilometer.agent [-] Config file: {'sources': [{'name': 'pollsters', 'interval': 120, 'meters': ['power.state', 'cpu', 'memory.usage', 'disk.*', 'network.*']}]} load_config /usr/lib/python3.9/site-packages/ceilometer/agent.py:64 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.126 12 DEBUG ceilometer.compute.virt.libvirt.utils [-] Connecting to libvirt: qemu:///system new_libvirt_connection /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/utils.py:93 Nov 27 04:35:44 localhost python3.9[240260]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/node_exporter/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236143.366615-1547-220702971452231/.source _original_basename=healthcheck follow=False checksum=e380c11c36804bfc65a818f2960cfa663daacfe5 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.469 12 DEBUG novaclient.v2.client [-] REQ: curl -g -i -X GET http://nova-internal.openstack.svc:8774/v2.1/flavors?is_public=None -H "Accept: application/json" -H "User-Agent: python-novaclient" -H "X-Auth-Token: {SHA256}6352d1980d2c0f4af048361bd60ef258afdfd692d327e4fc4d1e5ed695775434" -H "X-OpenStack-Nova-API-Version: 2.1" _http_log_request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:519 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.521 12 DEBUG novaclient.v2.client [-] RESP: [200] Connection: Keep-Alive Content-Length: 327 Content-Type: application/json Date: Thu, 27 Nov 2025 09:35:44 GMT Keep-Alive: timeout=5, max=100 OpenStack-API-Version: compute 2.1 Server: Apache Vary: OpenStack-API-Version,X-OpenStack-Nova-API-Version X-OpenStack-Nova-API-Version: 2.1 x-compute-request-id: req-a8fd0a72-c3cb-4bbc-9e58-dd82d9c0d505 x-openstack-request-id: req-a8fd0a72-c3cb-4bbc-9e58-dd82d9c0d505 _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:550 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.521 12 DEBUG novaclient.v2.client [-] RESP BODY: {"flavors": [{"id": "ad8d4a49-18f7-4954-9703-7480b3ae8896", "name": "m1.small", "links": [{"rel": "self", "href": "http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}, {"rel": "bookmark", "href": "http://nova-internal.openstack.svc:8774/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}]}]} _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:582 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.521 12 DEBUG novaclient.v2.client [-] GET call to compute for http://nova-internal.openstack.svc:8774/v2.1/flavors?is_public=None used request id req-a8fd0a72-c3cb-4bbc-9e58-dd82d9c0d505 request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:954 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.523 12 DEBUG novaclient.v2.client [-] REQ: curl -g -i -X GET http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896 -H "Accept: application/json" -H "User-Agent: python-novaclient" -H "X-Auth-Token: {SHA256}6352d1980d2c0f4af048361bd60ef258afdfd692d327e4fc4d1e5ed695775434" -H "X-OpenStack-Nova-API-Version: 2.1" _http_log_request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:519 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.543 12 DEBUG novaclient.v2.client [-] RESP: [200] Connection: Keep-Alive Content-Length: 494 Content-Type: application/json Date: Thu, 27 Nov 2025 09:35:44 GMT Keep-Alive: timeout=5, max=99 OpenStack-API-Version: compute 2.1 Server: Apache Vary: OpenStack-API-Version,X-OpenStack-Nova-API-Version X-OpenStack-Nova-API-Version: 2.1 x-compute-request-id: req-fdbd17a7-5630-443d-a146-4a2c616d9561 x-openstack-request-id: req-fdbd17a7-5630-443d-a146-4a2c616d9561 _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:550 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.543 12 DEBUG novaclient.v2.client [-] RESP BODY: {"flavor": {"id": "ad8d4a49-18f7-4954-9703-7480b3ae8896", "name": "m1.small", "ram": 512, "disk": 1, "swap": "", "OS-FLV-EXT-DATA:ephemeral": 1, "OS-FLV-DISABLED:disabled": false, "vcpus": 1, "os-flavor-access:is_public": true, "rxtx_factor": 1.0, "links": [{"rel": "self", "href": "http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}, {"rel": "bookmark", "href": "http://nova-internal.openstack.svc:8774/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896"}]}} _http_log_response /usr/lib/python3.9/site-packages/keystoneauth1/session.py:582 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.543 12 DEBUG novaclient.v2.client [-] GET call to compute for http://nova-internal.openstack.svc:8774/v2.1/flavors/ad8d4a49-18f7-4954-9703-7480b3ae8896 used request id req-fdbd17a7-5630-443d-a146-4a2c616d9561 request /usr/lib/python3.9/site-packages/keystoneauth1/session.py:954 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.545 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.545 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.583 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 498 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.584 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd34ce13a-ebbc-4509-a1bd-4fb93460b1e5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 498, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.545987', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '727bbbec-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '5dfb66dc0cc8bc7aec5b5bf359ecd96bcf22cc9581b78cda732c8ac5eef4cbc0'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.545987', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '727bd546-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '5f5e0f279a9cc312abbb0f5c4ddaec24959ccd05092292c8282be2fbe0e6236d'}]}, 'timestamp': '2025-11-27 09:35:44.585113', '_unique_id': '157fb46bf45d42f0b639eb988efdec49'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.593 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.596 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.601 12 DEBUG ceilometer.compute.virt.libvirt.inspector [-] No delta meter predecessor for a02f7f2f-d2cf-4612-b18a-8be435257201 / tapa2718872-3b inspect_vnics /usr/lib/python3.9/site-packages/ceilometer/compute/virt/libvirt/inspector.py:136 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.602 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5b57a5b4-e033-4af5-83e9-2233e53e0eb1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.597073', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '727e7ee0-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': '9aa7a4c25f0a65f768ffac5f921f581f8330b5d3d9b5448e88a51de45579925a'}]}, 'timestamp': '2025-11-27 09:35:44.602606', '_unique_id': '6e4cbb6a33fb4a74ab7b4eca394cd0c1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.603 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.604 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.605 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 11314 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '14972815-6547-45cc-8afa-33dbc8f51219', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 11314, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.605111', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '727ef668-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'c06d98161bd87365c5a506a99fd6be1e7646c747441dfb81f80db13d922dca1d'}]}, 'timestamp': '2025-11-27 09:35:44.605632', '_unique_id': '14c5bc8b17d7423e9462d63a338c2f9f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.606 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.607 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.607 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1b2b68a7-4b2e-478c-92a2-c5846901fc7c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.607820', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '727f6436-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'f6f3e4860e1b9f08077bacf183b12b4b7e4c3e42385bda729a73269d6e139dc8'}]}, 'timestamp': '2025-11-27 09:35:44.608485', '_unique_id': 'dd23033d5e924be59e041fcbbeef2702'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.609 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.611 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.611 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '95a80303-086b-46ff-9223-16d29f45a708', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.611183', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '727fe366-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'c5563cafc4c0644c9502ad82bf5878a7e20862689c69f3c204cdf10c513c73db'}]}, 'timestamp': '2025-11-27 09:35:44.611693', '_unique_id': 'ed0e0c3314e84fa2b73a57c4bd5ea152'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.612 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.614 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.614 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1064 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.614 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 222 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0c35e765-c6b1-455d-b85f-93c3119a0c08', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1064, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.614190', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728058b4-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '0b2163c5daa2b530104c94755e13996b851ec9ac5a4aeff4cec2d9c9071e2807'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 222, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.614190', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '72806ade-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '70b7a197bf3538aa16b95c44751c1c09720d4a277e1720647a12e021356b4642'}]}, 'timestamp': '2025-11-27 09:35:44.615090', '_unique_id': '9870c4cc1e5841e087052246f5cf6d32'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.616 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.617 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.617 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '534f05f5-15de-4070-a75b-d83a2b869dab', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.617329', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '7280d4a6-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': '989989908c6d3dcfc6685ea5abbed158e31da52cbe2a3c531434108f0f2feca1'}]}, 'timestamp': '2025-11-27 09:35:44.617833', '_unique_id': '7de90f3784fc4e449a7f16bac36afc36'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.618 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.619 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.619 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 130 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'fe9868e0-caef-4604-9c21-8566f84ebf9a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 130, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.619949', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '72813b1c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': '2d463d5b81b6e38950235c290ae90ed0f75c683ad3ff38931965a8919efa8b12'}]}, 'timestamp': '2025-11-27 09:35:44.620457', '_unique_id': '633abab2be4f4e099251d52f985655de'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.621 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.622 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.622 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 974916304 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.623 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 177343418 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '751b0658-5137-4230-8222-99254db6e6d8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 974916304, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.622718', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '7281a5b6-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': 'cf2193dff239d776b229d9353a43df6b00dd5e6c2d9589101d683ee4b3859856'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 177343418, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.622718', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '7281b63c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '327e4458bc1f016ad5280c7cbb8895ca682204597b186ac960cf75c3d5d46ada'}]}, 'timestamp': '2025-11-27 09:35:44.623606', '_unique_id': 'd65cd10d0f0246e997402be7809bd53a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.624 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.625 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.648 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 52.46875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f84dd36d-4529-4d7d-8f3b-682890ab1bf7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 52.46875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:35:44.625804', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '7285a0da-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.836420985, 'message_signature': '966e0f28c1bc83f16e5a1b07935596d0c7f0e94f1340912b5b1dff7390b47f9a'}]}, 'timestamp': '2025-11-27 09:35:44.649257', '_unique_id': 'b90a871d0d3e4b52843af1c21974daaf'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.650 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.651 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.rate in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.651 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for IncomingBytesRatePollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.651 12 ERROR ceilometer.polling.manager [-] Prevent pollster network.incoming.bytes.rate from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.652 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.iops in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.652 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for PerDeviceDiskIOPSPollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.652 12 ERROR ceilometer.polling.manager [-] Prevent pollster disk.device.iops from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.652 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.rate in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.653 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for OutgoingBytesRatePollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.653 12 ERROR ceilometer.polling.manager [-] Prevent pollster network.outgoing.bytes.rate from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.653 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.653 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 29130240 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.654 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 4300800 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7884cd2f-67cb-49c4-adfc-69748b9c582d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 29130240, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.653569', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '72865b2e-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '4e9ec27465825aeaf24290f7f368a0e0833e364e27fbbf4b6941e444f2ff2c40'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 4300800, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.653569', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '72866b82-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': 'f50f7b91f208ae1e5bbccb4b2b3ee32735baa06f3371628440ae203cfafd3447'}]}, 'timestamp': '2025-11-27 09:35:44.654426', '_unique_id': 'eb78a119064e4592b6975bb67562f51a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.655 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.656 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.656 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a3c66dcc-3635-410a-9b86-131f8db2c262', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.656659', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '7286d504-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': '3af85338fa4b6ee0dedda56fc65cddb3c1d8da30c91df72f51d739320037104a'}]}, 'timestamp': '2025-11-27 09:35:44.657154', '_unique_id': 'f4855e94a25446da87c3b5ffc814a3f5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.658 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.659 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.659 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 73908224 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.659 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3e361c64-627f-4116-8a13-f8f4630af21e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 73908224, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.659259', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728738dc-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '326d40173bcd571d69f9613f4aeb75705ef83fbe9486da92b3e64d01dd659f06'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.659259', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '72874a52-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '99d05ca635c2509f3a6d394d48b451881c3518f073bb018333751959162e4dec'}]}, 'timestamp': '2025-11-27 09:35:44.660124', '_unique_id': '0fbd8407f22d4a8ab91a674b7ce59237'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.661 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.662 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.662 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 48280000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0f3da699-9317-47e8-94c7-2c35013da453', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 48280000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:35:44.662282', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '7287aef2-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.836420985, 'message_signature': '679e50e21ac496a7d68bc1ad70825512ea0df9a0043c0711b9d7bb78eb240e49'}]}, 'timestamp': '2025-11-27 09:35:44.662746', '_unique_id': '1fb3a8d5eb0145b5af24b4aa808f91c8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.663 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.664 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.664 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '07b07d9d-da16-49ae-9730-bab6fa12633b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.664856', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '7288152c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'a5d4c484646b357138c8c3a1391347d3cfd25bb46d5b5f626d2504b27c41b548'}]}, 'timestamp': '2025-11-27 09:35:44.665349', '_unique_id': '7966be1cf9a2409198ee7da69b2bc08d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.666 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.667 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.680 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.681 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '87e1be9b-dcc8-4300-b082-9940127bf510', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.667439', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728a842e-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': 'c6dd8828f4c8c1af17c7f4e173dcb077d61e56e9c341fc62f27ddcd8bf870f67'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.667439', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '728a94d2-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': '990be779d84edd5d0b13939221004075fb94dcb0ca08ff7c74e6fb037508ce31'}]}, 'timestamp': '2025-11-27 09:35:44.681729', '_unique_id': '5a93f9bfb1f349fcbed5f79cf59d1666'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.682 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.683 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.latency in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.683 12 DEBUG ceilometer.compute.pollsters [-] LibvirtInspector does not provide data for PerDeviceDiskLatencyPollster get_samples /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:163 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.684 12 ERROR ceilometer.polling.manager [-] Prevent pollster disk.device.latency from polling [] on source pollsters anymore!: ceilometer.polling.plugin_base.PollsterPermanentError: [] Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.684 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.684 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.684 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3d76b0ad-eb35-4681-9359-2bc1406bd146', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.684492', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728b143e-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': '10e526b976886a80a8bff32a5ad190cedf05d284b04b3a267cadb15519026aef'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.684492', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '728b253c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': 'ea61317dfeb8dbb87a022a0265a687056f96c963a00e898c945847166b12dcd2'}]}, 'timestamp': '2025-11-27 09:35:44.685395', '_unique_id': 'c5c17fe4846a4684b4f9f819054d73cb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.686 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.687 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.687 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 83 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ccff6b17-aec2-4142-8325-11a820d45faf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 83, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.687593', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '728b8c0c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'cea9176f75343fcf73115a1c7afd8883b0c0485c2600722f67dfa3011e0b29c5'}]}, 'timestamp': '2025-11-27 09:35:44.688051', '_unique_id': 'b315ba91c77c4314a1809377349c6b2c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.688 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.690 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.690 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 8825 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4e52c4aa-8dd6-423f-8fd1-78c725e0158d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 8825, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:35:44.690176', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '728bf0c0-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.785301938, 'message_signature': 'ae9abee758fc1fd22dedf08fcfbb84ceeec38174296dba44cc69a6131f61831f'}]}, 'timestamp': '2025-11-27 09:35:44.690667', '_unique_id': 'b54da00448dc422da5555af50bc3de9d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.691 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.692 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.692 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 203265086 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.693 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 25604091 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e4fb41e4-7800-4bdd-b60a-8b1f32e335b6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 203265086, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.692756', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728c563c-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': '210d720be59811cc7641d50a275b25ea336ede71e9295f0bb33838659f5e42c7'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 25604091, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.692756', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '728c6686-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.73414836, 'message_signature': 'a05459cad90858c955d4a8271f50e4a4a3390e8e3f00f1ba7144fe65eecc263b'}]}, 'timestamp': '2025-11-27 09:35:44.693646', '_unique_id': '32d1d3bbc5ea4e449e19a245b26fcbed'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.694 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.695 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.696 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.696 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '41f9125a-ca71-4835-8aab-9721177e193a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:35:44.695965', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '728cd2b0-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': 'e6c7db3b5b7442e976ea30946f9e62c7f1f60942448ea96561b6f79fb38e7a91'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:35:44.695965', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '728ce3fe-cb74-11f0-8166-fa163e6bfee1', 'monotonic_time': 10374.855640896, 'message_signature': '65bf1033a4dc8790fdb4946465e7cb312459528b98986f1d1a08614c55177c37'}]}, 'timestamp': '2025-11-27 09:35:44.696828', '_unique_id': '38b7abe0411c4feb865c6ff4f26d6ed2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:35:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:35:44.697 12 ERROR oslo_messaging.notify.messaging Nov 27 04:35:45 localhost python3.9[240370]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/telemetry config_pattern=node_exporter.json debug=False Nov 27 04:35:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=41555 DF PROTO=TCP SPT=56308 DPT=9102 SEQ=1333734345 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129590690000000001030307) Nov 27 04:35:45 localhost nova_compute[232618]: 2025-11-27 09:35:45.698 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:46 localhost python3.9[240480]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:35:47 localhost python3[240590]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/telemetry config_id=edpm config_overrides={} config_patterns=node_exporter.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:35:47 localhost podman[240629]: Nov 27 04:35:47 localhost podman[240629]: 2025-11-27 09:35:47.291283163 +0000 UTC m=+0.079178466 container create d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, maintainer=The Prometheus Authors , config_id=edpm, container_name=node_exporter, managed_by=edpm_ansible) Nov 27 04:35:47 localhost podman[240629]: 2025-11-27 09:35:47.246022384 +0000 UTC m=+0.033917727 image pull quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c Nov 27 04:35:47 localhost python3[240590]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name node_exporter --conmon-pidfile /run/node_exporter.pid --env OS_ENDPOINT_TYPE=internal --healthcheck-command /openstack/healthcheck node_exporter --label config_id=edpm --label container_name=node_exporter --label managed_by=edpm_ansible --label config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']} --log-driver journald --log-level info --network host --privileged=True --publish 9100:9100 --user root --volume /var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw --volume /var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c --web.disable-exporter-metrics --collector.systemd --collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\.service --no-collector.dmi --no-collector.entropy --no-collector.thermal_zone --no-collector.time --no-collector.timex --no-collector.uname --no-collector.stat --no-collector.hwmon --no-collector.os --no-collector.selinux --no-collector.textfile --no-collector.powersupplyclass --no-collector.pressure --no-collector.rapl Nov 27 04:35:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31488 DF PROTO=TCP SPT=35094 DPT=9100 SEQ=934839684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295989A0000000001030307) Nov 27 04:35:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:35:48 localhost systemd[1]: tmp-crun.PF1ObM.mount: Deactivated successfully. Nov 27 04:35:48 localhost podman[240771]: 2025-11-27 09:35:48.017885505 +0000 UTC m=+0.095705666 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, container_name=ovn_controller) Nov 27 04:35:48 localhost podman[240771]: 2025-11-27 09:35:48.085950564 +0000 UTC m=+0.163770735 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, container_name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:35:48 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:35:48 localhost python3.9[240784]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:35:49 localhost python3.9[240908]: ansible-file Invoked with path=/etc/systemd/system/edpm_node_exporter.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:49 localhost nova_compute[232618]: 2025-11-27 09:35:49.039 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:49 localhost python3.9[241017]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236149.0937386-1705-118233534396531/source dest=/etc/systemd/system/edpm_node_exporter.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:35:50 localhost python3.9[241072]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:35:50 localhost systemd[1]: Reloading. Nov 27 04:35:50 localhost systemd-rc-local-generator[241095]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:35:50 localhost systemd-sysv-generator[241098]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:50 localhost nova_compute[232618]: 2025-11-27 09:35:50.725 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:51 localhost python3.9[241163]: ansible-systemd Invoked with state=restarted name=edpm_node_exporter.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:35:51 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31489 DF PROTO=TCP SPT=35094 DPT=9100 SEQ=934839684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295A85B0000000001030307) Nov 27 04:35:52 localhost systemd[1]: Reloading. Nov 27 04:35:52 localhost systemd-sysv-generator[241190]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:35:52 localhost systemd-rc-local-generator[241186]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:35:52 localhost systemd[1]: Starting node_exporter container... Nov 27 04:35:52 localhost systemd[1]: Started libcrun container. Nov 27 04:35:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:35:52 localhost podman[241203]: 2025-11-27 09:35:52.741236555 +0000 UTC m=+0.136206402 container init d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.759Z caller=node_exporter.go:180 level=info msg="Starting node_exporter" version="(version=1.5.0, branch=HEAD, revision=1b48970ffcf5630534fb00bb0687d73c66d1c959)" Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.759Z caller=node_exporter.go:181 level=info msg="Build context" build_context="(go=go1.19.3, user=root@6e7732a7b81b, date=20221129-18:59:09)" Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.759Z caller=node_exporter.go:183 level=warn msg="Node Exporter is running as root user. This exporter is designed to run as unprivileged user, root is not required." Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.759Z caller=diskstats_common.go:111 level=info collector=diskstats msg="Parsed flag --collector.diskstats.device-exclude" flag=^(ram|loop|fd|(h|s|v|xv)d[a-z]|nvme\d+n\d+p)\d+$ Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.759Z caller=diskstats_linux.go:264 level=error collector=diskstats msg="Failed to open directory, disabling udev device properties" path=/run/udev/data Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.760Z caller=systemd_linux.go:152 level=info collector=systemd msg="Parsed flag --collector.systemd.unit-include" flag=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\.service Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.760Z caller=systemd_linux.go:154 level=info collector=systemd msg="Parsed flag --collector.systemd.unit-exclude" flag=.+\.(automount|device|mount|scope|slice) Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=filesystem_common.go:111 level=info collector=filesystem msg="Parsed flag --collector.filesystem.mount-points-exclude" flag=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/containers/storage/.+)($|/) Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=filesystem_common.go:113 level=info collector=filesystem msg="Parsed flag --collector.filesystem.fs-types-exclude" flag=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:110 level=info msg="Enabled collectors" Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=arp Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=bcache Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=bonding Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=btrfs Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=conntrack Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=cpu Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=cpufreq Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=diskstats Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=edac Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=fibrechannel Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=filefd Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=filesystem Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=infiniband Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=ipvs Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=loadavg Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=mdadm Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=meminfo Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=netclass Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=netdev Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=netstat Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=nfs Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=nfsd Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=nvme Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=schedstat Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=sockstat Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=softnet Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=systemd Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=tapestats Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=udp_queues Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=vmstat Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=xfs Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.761Z caller=node_exporter.go:117 level=info collector=zfs Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.762Z caller=tls_config.go:232 level=info msg="Listening on" address=[::]:9100 Nov 27 04:35:52 localhost node_exporter[241218]: ts=2025-11-27T09:35:52.763Z caller=tls_config.go:235 level=info msg="TLS is disabled." http2=false address=[::]:9100 Nov 27 04:35:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:35:52 localhost podman[241203]: 2025-11-27 09:35:52.776957965 +0000 UTC m=+0.171927792 container start d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:35:52 localhost podman[241203]: node_exporter Nov 27 04:35:52 localhost systemd[1]: Started node_exporter container. Nov 27 04:35:52 localhost podman[241227]: 2025-11-27 09:35:52.861645418 +0000 UTC m=+0.078197557 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=starting, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:35:52 localhost podman[241227]: 2025-11-27 09:35:52.905879755 +0000 UTC m=+0.122431814 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:35:52 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:35:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:35:53 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44162 DF PROTO=TCP SPT=55642 DPT=9101 SEQ=304980020 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295B05A0000000001030307) Nov 27 04:35:53 localhost systemd[1]: tmp-crun.Cleq2l.mount: Deactivated successfully. Nov 27 04:35:53 localhost podman[241305]: 2025-11-27 09:35:53.757011694 +0000 UTC m=+0.094901573 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125) Nov 27 04:35:53 localhost podman[241305]: 2025-11-27 09:35:53.761943923 +0000 UTC m=+0.099833782 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:35:53 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:35:54 localhost nova_compute[232618]: 2025-11-27 09:35:54.071 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:54 localhost python3.9[241377]: ansible-ansible.builtin.systemd Invoked with name=edpm_node_exporter.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:35:54 localhost systemd[1]: Stopping node_exporter container... Nov 27 04:35:54 localhost systemd[1]: libpod-d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.scope: Deactivated successfully. Nov 27 04:35:54 localhost podman[241381]: 2025-11-27 09:35:54.271819418 +0000 UTC m=+0.062561384 container died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:35:54 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.timer: Deactivated successfully. Nov 27 04:35:54 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:35:54 localhost podman[241381]: 2025-11-27 09:35:54.316561101 +0000 UTC m=+0.107303067 container cleanup d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:35:54 localhost podman[241381]: node_exporter Nov 27 04:35:54 localhost systemd[1]: edpm_node_exporter.service: Main process exited, code=exited, status=2/INVALIDARGUMENT Nov 27 04:35:54 localhost podman[241408]: 2025-11-27 09:35:54.414821874 +0000 UTC m=+0.067556775 container cleanup d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:35:54 localhost podman[241408]: node_exporter Nov 27 04:35:54 localhost systemd[1]: edpm_node_exporter.service: Failed with result 'exit-code'. Nov 27 04:35:54 localhost systemd[1]: Stopped node_exporter container. Nov 27 04:35:54 localhost systemd[1]: Starting node_exporter container... Nov 27 04:35:54 localhost systemd[1]: Started libcrun container. Nov 27 04:35:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:35:54 localhost podman[241421]: 2025-11-27 09:35:54.577876026 +0000 UTC m=+0.128245841 container init d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.591Z caller=node_exporter.go:180 level=info msg="Starting node_exporter" version="(version=1.5.0, branch=HEAD, revision=1b48970ffcf5630534fb00bb0687d73c66d1c959)" Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.591Z caller=node_exporter.go:181 level=info msg="Build context" build_context="(go=go1.19.3, user=root@6e7732a7b81b, date=20221129-18:59:09)" Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.591Z caller=node_exporter.go:183 level=warn msg="Node Exporter is running as root user. This exporter is designed to run as unprivileged user, root is not required." Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.592Z caller=diskstats_common.go:111 level=info collector=diskstats msg="Parsed flag --collector.diskstats.device-exclude" flag=^(ram|loop|fd|(h|s|v|xv)d[a-z]|nvme\d+n\d+p)\d+$ Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.592Z caller=diskstats_linux.go:264 level=error collector=diskstats msg="Failed to open directory, disabling udev device properties" path=/run/udev/data Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.592Z caller=filesystem_common.go:111 level=info collector=filesystem msg="Parsed flag --collector.filesystem.mount-points-exclude" flag=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/containers/storage/.+)($|/) Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.592Z caller=filesystem_common.go:113 level=info collector=filesystem msg="Parsed flag --collector.filesystem.fs-types-exclude" flag=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=systemd_linux.go:152 level=info collector=systemd msg="Parsed flag --collector.systemd.unit-include" flag=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\.service Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=systemd_linux.go:154 level=info collector=systemd msg="Parsed flag --collector.systemd.unit-exclude" flag=.+\.(automount|device|mount|scope|slice) Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:110 level=info msg="Enabled collectors" Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=arp Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=bcache Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=bonding Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=btrfs Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=conntrack Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=cpu Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=cpufreq Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=diskstats Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=edac Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=fibrechannel Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=filefd Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=filesystem Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=infiniband Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=ipvs Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=loadavg Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=mdadm Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=meminfo Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=netclass Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=netdev Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=netstat Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=nfs Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=nfsd Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=nvme Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=schedstat Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=sockstat Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=softnet Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=systemd Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=tapestats Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=udp_queues Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=vmstat Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=xfs Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.593Z caller=node_exporter.go:117 level=info collector=zfs Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.594Z caller=tls_config.go:232 level=info msg="Listening on" address=[::]:9100 Nov 27 04:35:54 localhost node_exporter[241435]: ts=2025-11-27T09:35:54.594Z caller=tls_config.go:235 level=info msg="TLS is disabled." http2=false address=[::]:9100 Nov 27 04:35:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:35:54 localhost podman[241421]: 2025-11-27 09:35:54.612644399 +0000 UTC m=+0.163014184 container start d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:35:54 localhost podman[241421]: node_exporter Nov 27 04:35:54 localhost systemd[1]: Started node_exporter container. Nov 27 04:35:54 localhost systemd[1]: tmp-crun.zHNz0z.mount: Deactivated successfully. Nov 27 04:35:54 localhost systemd[1]: tmp-crun.pxolV7.mount: Deactivated successfully. Nov 27 04:35:54 localhost podman[241444]: 2025-11-27 09:35:54.703724173 +0000 UTC m=+0.086544449 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=starting, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:35:54 localhost podman[241444]: 2025-11-27 09:35:54.717731507 +0000 UTC m=+0.100551813 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:35:54 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:35:55 localhost nova_compute[232618]: 2025-11-27 09:35:55.761 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:56 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=55651 DF PROTO=TCP SPT=60286 DPT=9101 SEQ=1365417040 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295BBDA0000000001030307) Nov 27 04:35:56 localhost python3.9[241575]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/podman_exporter/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:35:57 localhost python3.9[241663]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/podman_exporter/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236156.3569045-1802-179169937060329/.source _original_basename=healthcheck follow=False checksum=e380c11c36804bfc65a818f2960cfa663daacfe5 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:35:58 localhost python3.9[241773]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/telemetry config_pattern=podman_exporter.json debug=False Nov 27 04:35:59 localhost nova_compute[232618]: 2025-11-27 09:35:59.107 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:35:59 localhost python3.9[241883]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:35:59 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31490 DF PROTO=TCP SPT=35094 DPT=9100 SEQ=934839684 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295C7DA0000000001030307) Nov 27 04:36:00 localhost nova_compute[232618]: 2025-11-27 09:36:00.806 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:01 localhost python3[241993]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/telemetry config_id=edpm config_overrides={} config_patterns=podman_exporter.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:36:02 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=38987 DF PROTO=TCP SPT=55658 DPT=9105 SEQ=1023967915 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295D41A0000000001030307) Nov 27 04:36:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:36:04 localhost podman[242021]: 2025-11-27 09:36:04.064943117 +0000 UTC m=+0.158762752 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=multipathd, container_name=multipathd, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:36:04 localhost podman[242021]: 2025-11-27 09:36:04.080799911 +0000 UTC m=+0.174619586 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 04:36:04 localhost nova_compute[232618]: 2025-11-27 09:36:04.134 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:04 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:36:04 localhost podman[242008]: 2025-11-27 09:36:01.772018236 +0000 UTC m=+0.044188240 image pull quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd Nov 27 04:36:04 localhost podman[242100]: Nov 27 04:36:04 localhost podman[242100]: 2025-11-27 09:36:04.747154112 +0000 UTC m=+0.094545631 container create a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, maintainer=Navid Yaghoobi ) Nov 27 04:36:04 localhost podman[242100]: 2025-11-27 09:36:04.69999179 +0000 UTC m=+0.047383359 image pull quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd Nov 27 04:36:04 localhost python3[241993]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: podman create --name podman_exporter --conmon-pidfile /run/podman_exporter.pid --env OS_ENDPOINT_TYPE=internal --env CONTAINER_HOST=unix:///run/podman/podman.sock --healthcheck-command /openstack/healthcheck podman_exporter --label config_id=edpm --label container_name=podman_exporter --label managed_by=edpm_ansible --label config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']} --log-driver journald --log-level info --network host --privileged=True --publish 9882:9882 --user root --volume /run/podman/podman.sock:/run/podman/podman.sock:rw,z --volume /var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd Nov 27 04:36:05 localhost nova_compute[232618]: 2025-11-27 09:36:05.841 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:06 localhost python3.9[242245]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:36:06 localhost python3.9[242357]: ansible-file Invoked with path=/etc/systemd/system/edpm_podman_exporter.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:36:07 localhost python3.9[242466]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236166.9150217-1960-54231277296025/source dest=/etc/systemd/system/edpm_podman_exporter.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:36:07 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=44165 DF PROTO=TCP SPT=55642 DPT=9101 SEQ=304980020 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295E7DA0000000001030307) Nov 27 04:36:08 localhost python3.9[242521]: ansible-systemd Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:36:08 localhost systemd[1]: Reloading. Nov 27 04:36:08 localhost systemd-sysv-generator[242550]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:36:08 localhost systemd-rc-local-generator[242545]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:08 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31966 DF PROTO=TCP SPT=55800 DPT=9882 SEQ=3854079338 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295EB070000000001030307) Nov 27 04:36:09 localhost python3.9[242612]: ansible-systemd Invoked with state=restarted name=edpm_podman_exporter.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:36:09 localhost nova_compute[232618]: 2025-11-27 09:36:09.171 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:09 localhost systemd[1]: Reloading. Nov 27 04:36:09 localhost systemd-rc-local-generator[242638]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:36:09 localhost systemd-sysv-generator[242644]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:36:09 localhost systemd[1]: Starting podman_exporter container... Nov 27 04:36:09 localhost systemd[1]: Started libcrun container. Nov 27 04:36:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:36:09 localhost podman[242653]: 2025-11-27 09:36:09.737940806 +0000 UTC m=+0.155547653 container init a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:36:09 localhost podman_exporter[242667]: ts=2025-11-27T09:36:09.759Z caller=exporter.go:68 level=info msg="Starting podman-prometheus-exporter" version="(version=1.10.1, branch=HEAD, revision=1)" Nov 27 04:36:09 localhost podman_exporter[242667]: ts=2025-11-27T09:36:09.759Z caller=exporter.go:69 level=info msg=metrics enhanced=false Nov 27 04:36:09 localhost podman_exporter[242667]: ts=2025-11-27T09:36:09.759Z caller=handler.go:94 level=info msg="enabled collectors" Nov 27 04:36:09 localhost podman_exporter[242667]: ts=2025-11-27T09:36:09.760Z caller=handler.go:105 level=info collector=container Nov 27 04:36:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:36:09 localhost podman[242653]: 2025-11-27 09:36:09.775277072 +0000 UTC m=+0.192883869 container start a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:36:09 localhost podman[242653]: podman_exporter Nov 27 04:36:09 localhost systemd[1]: Starting Podman API Service... Nov 27 04:36:09 localhost systemd[1]: Started Podman API Service. Nov 27 04:36:09 localhost systemd[1]: Started podman_exporter container. Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="/usr/bin/podman filtering at log level info" Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="Not using native diff for overlay, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="Setting parallel job count to 25" Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="Using systemd socket activation to determine API endpoint" Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="API service listening on \"/run/podman/podman.sock\". URI: \"/run/podman/podman.sock\"" Nov 27 04:36:09 localhost podman[242678]: @ - - [27/Nov/2025:09:36:09 +0000] "GET /v4.9.3/libpod/_ping HTTP/1.1" 200 2 "" "Go-http-client/1.1" Nov 27 04:36:09 localhost podman[242678]: time="2025-11-27T09:36:09Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:36:09 localhost podman[242677]: 2025-11-27 09:36:09.913680603 +0000 UTC m=+0.128810593 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=starting, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:36:09 localhost podman[242677]: 2025-11-27 09:36:09.922119729 +0000 UTC m=+0.137249779 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:36:09 localhost podman[242677]: unhealthy Nov 27 04:36:10 localhost systemd[1]: tmp-crun.Jt4uQe.mount: Deactivated successfully. Nov 27 04:36:10 localhost nova_compute[232618]: 2025-11-27 09:36:10.845 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:11 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:11 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:36:11 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Failed with result 'exit-code'. Nov 27 04:36:11 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:11 localhost python3.9[242825]: ansible-ansible.builtin.systemd Invoked with name=edpm_podman_exporter.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:36:11 localhost systemd[1]: Stopping podman_exporter container... Nov 27 04:36:11 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=31968 DF PROTO=TCP SPT=55800 DPT=9882 SEQ=3854079338 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1295F71A0000000001030307) Nov 27 04:36:11 localhost podman[242678]: @ - - [27/Nov/2025:09:36:09 +0000] "GET /v4.9.3/libpod/events?filters=%7B%7D&since=&stream=true&until= HTTP/1.1" 200 2790 "" "Go-http-client/1.1" Nov 27 04:36:11 localhost systemd[1]: libpod-a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.scope: Deactivated successfully. Nov 27 04:36:11 localhost podman[242829]: 2025-11-27 09:36:11.908580608 +0000 UTC m=+0.069400487 container died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:36:11 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.timer: Deactivated successfully. Nov 27 04:36:11 localhost systemd[1]: Stopped /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:36:11 localhost systemd[1]: tmp-crun.3j9Ssp.mount: Deactivated successfully. Nov 27 04:36:12 localhost podman[242829]: 2025-11-27 09:36:12.43801853 +0000 UTC m=+0.598838429 container cleanup a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:36:12 localhost podman[242829]: podman_exporter Nov 27 04:36:12 localhost podman[242841]: 2025-11-27 09:36:12.445297274 +0000 UTC m=+0.532725845 container cleanup a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:36:12 localhost systemd[1]: var-lib-containers-storage-overlay-f54f2c991c5dc053c987f5ebdf34aeb7b43df9c2b2cbbe0d1ad204b939241a5e-merged.mount: Deactivated successfully. Nov 27 04:36:12 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e-userdata-shm.mount: Deactivated successfully. Nov 27 04:36:13 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:13 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:13 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:36:13 localhost systemd[1]: edpm_podman_exporter.service: Main process exited, code=exited, status=2/INVALIDARGUMENT Nov 27 04:36:13 localhost podman[242856]: 2025-11-27 09:36:13.645504171 +0000 UTC m=+0.085951101 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=starting, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, config_id=edpm) Nov 27 04:36:13 localhost podman[242856]: 2025-11-27 09:36:13.679498554 +0000 UTC m=+0.119945504 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute) Nov 27 04:36:13 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:13 localhost podman[242856]: unhealthy Nov 27 04:36:13 localhost podman[242867]: 2025-11-27 09:36:13.695391559 +0000 UTC m=+0.071348651 container cleanup a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:36:13 localhost podman[242867]: podman_exporter Nov 27 04:36:13 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:13 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:36:13 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Failed with result 'exit-code'. Nov 27 04:36:13 localhost systemd[1]: edpm_podman_exporter.service: Failed with result 'exit-code'. Nov 27 04:36:13 localhost systemd[1]: Stopped podman_exporter container. Nov 27 04:36:13 localhost systemd[1]: Starting podman_exporter container... Nov 27 04:36:14 localhost nova_compute[232618]: 2025-11-27 09:36:14.173 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:14 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:14 localhost systemd[1]: Started libcrun container. Nov 27 04:36:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:36:15 localhost podman[242885]: 2025-11-27 09:36:15.028219554 +0000 UTC m=+1.290410576 container init a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:36:15 localhost podman_exporter[242900]: ts=2025-11-27T09:36:15.048Z caller=exporter.go:68 level=info msg="Starting podman-prometheus-exporter" version="(version=1.10.1, branch=HEAD, revision=1)" Nov 27 04:36:15 localhost podman_exporter[242900]: ts=2025-11-27T09:36:15.048Z caller=exporter.go:69 level=info msg=metrics enhanced=false Nov 27 04:36:15 localhost podman_exporter[242900]: ts=2025-11-27T09:36:15.048Z caller=handler.go:94 level=info msg="enabled collectors" Nov 27 04:36:15 localhost podman_exporter[242900]: ts=2025-11-27T09:36:15.048Z caller=handler.go:105 level=info collector=container Nov 27 04:36:15 localhost podman[242678]: @ - - [27/Nov/2025:09:36:15 +0000] "GET /v4.9.3/libpod/_ping HTTP/1.1" 200 2 "" "Go-http-client/1.1" Nov 27 04:36:15 localhost podman[242678]: time="2025-11-27T09:36:15Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:36:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:36:15 localhost podman[242885]: 2025-11-27 09:36:15.077606249 +0000 UTC m=+1.339797261 container start a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:36:15 localhost podman[242885]: podman_exporter Nov 27 04:36:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36657 DF PROTO=TCP SPT=46168 DPT=9102 SEQ=722605955 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129605980000000001030307) Nov 27 04:36:15 localhost systemd[1]: tmp-crun.JhsW8t.mount: Deactivated successfully. Nov 27 04:36:15 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:15 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:15 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:15 localhost nova_compute[232618]: 2025-11-27 09:36:15.847 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:16 localhost systemd[1]: Started podman_exporter container. Nov 27 04:36:16 localhost podman[242911]: 2025-11-27 09:36:16.266080497 +0000 UTC m=+1.185647170 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=starting, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:36:16 localhost podman[242911]: 2025-11-27 09:36:16.306865451 +0000 UTC m=+1.226432114 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:36:16 localhost podman[242911]: unhealthy Nov 27 04:36:16 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:16 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:16 localhost systemd[1]: var-lib-containers-storage-overlay-c892fd6b7d17c3244e97732d72b83cd3d1a569af20da04450edaf25f54095ce6-merged.mount: Deactivated successfully. Nov 27 04:36:16 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:16 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:17 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:36:17 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Failed with result 'exit-code'. Nov 27 04:36:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48600 DF PROTO=TCP SPT=56462 DPT=9100 SEQ=2495205666 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12960DDA0000000001030307) Nov 27 04:36:18 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:18 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:36:18 localhost systemd[1]: var-lib-containers-storage-overlay-706e7dad99f72ad0a2bbd0167294cca4c86e34b8636b5f5f9b7b80818f253353-merged.mount: Deactivated successfully. Nov 27 04:36:18 localhost podman[242951]: 2025-11-27 09:36:18.20881071 +0000 UTC m=+0.081855586 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:36:18 localhost podman[242951]: 2025-11-27 09:36:18.241713332 +0000 UTC m=+0.114758248 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller) Nov 27 04:36:18 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:18 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:19 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:19 localhost nova_compute[232618]: 2025-11-27 09:36:19.210 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:19 localhost python3.9[243069]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/openstack_network_exporter/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:36:19 localhost systemd[1]: var-lib-containers-storage-overlay-f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048-merged.mount: Deactivated successfully. Nov 27 04:36:19 localhost python3.9[243157]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/healthchecks/openstack_network_exporter/ group=zuul mode=0700 owner=zuul setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236178.8833475-2057-90374869626558/.source _original_basename=healthcheck follow=False checksum=e380c11c36804bfc65a818f2960cfa663daacfe5 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None seuser=None serole=None selevel=None attributes=None Nov 27 04:36:20 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:20 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:20 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:36:20 localhost nova_compute[232618]: 2025-11-27 09:36:20.852 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:20 localhost systemd[1]: var-lib-containers-storage-overlay-55d5530fe8468c8c9907e0aa1de030811941604fa5f46de3db6dc15ec40906dd-merged.mount: Deactivated successfully. Nov 27 04:36:20 localhost python3.9[243267]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/telemetry config_pattern=openstack_network_exporter.json debug=False Nov 27 04:36:21 localhost systemd[1]: var-lib-containers-storage-overlay-ae0ebe7656e29542866ff018f5be9a3d02c88268a65814cf045e1b6c30ffd352-merged.mount: Deactivated successfully. Nov 27 04:36:21 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48601 DF PROTO=TCP SPT=56462 DPT=9100 SEQ=2495205666 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12961D9B0000000001030307) Nov 27 04:36:21 localhost python3.9[243377]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:36:22 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:22 localhost systemd[1]: var-lib-containers-storage-overlay-f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048-merged.mount: Deactivated successfully. Nov 27 04:36:22 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:22 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:22 localhost python3[243487]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/telemetry config_id=edpm config_overrides={} config_patterns=openstack_network_exporter.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:36:23 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:23 localhost systemd[1]: var-lib-containers-storage-overlay-c892fd6b7d17c3244e97732d72b83cd3d1a569af20da04450edaf25f54095ce6-merged.mount: Deactivated successfully. Nov 27 04:36:23 localhost systemd[1]: var-lib-containers-storage-overlay-c892fd6b7d17c3244e97732d72b83cd3d1a569af20da04450edaf25f54095ce6-merged.mount: Deactivated successfully. Nov 27 04:36:23 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4542 DF PROTO=TCP SPT=35184 DPT=9101 SEQ=2397825623 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1296255B0000000001030307) Nov 27 04:36:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:36:23 localhost systemd[1]: var-lib-containers-storage-overlay-f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a-merged.mount: Deactivated successfully. Nov 27 04:36:24 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:24 localhost nova_compute[232618]: 2025-11-27 09:36:24.236 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:24 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:36:24 localhost podman[243525]: 2025-11-27 09:36:24.995838229 +0000 UTC m=+0.090085357 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:36:25 localhost podman[243525]: 2025-11-27 09:36:25.007469225 +0000 UTC m=+0.101716353 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:36:25 localhost systemd[1]: var-lib-containers-storage-overlay-06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66-merged.mount: Deactivated successfully. Nov 27 04:36:25 localhost systemd[1]: var-lib-containers-storage-overlay-f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a-merged.mount: Deactivated successfully. Nov 27 04:36:25 localhost systemd[1]: var-lib-containers-storage-overlay-a802e2c2182c5081dae453e00ae55ca652c01124f4ff691b910ec76e11c97f5a-merged.mount: Deactivated successfully. Nov 27 04:36:25 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:36:25 localhost nova_compute[232618]: 2025-11-27 09:36:25.856 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:26 localhost systemd[1]: var-lib-containers-storage-overlay-f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a-merged.mount: Deactivated successfully. Nov 27 04:36:26 localhost systemd[1]: var-lib-containers-storage-overlay-706e7dad99f72ad0a2bbd0167294cca4c86e34b8636b5f5f9b7b80818f253353-merged.mount: Deactivated successfully. Nov 27 04:36:26 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36353 DF PROTO=TCP SPT=42340 DPT=9105 SEQ=3673889381 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1296319A0000000001030307) Nov 27 04:36:27 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:29 localhost nova_compute[232618]: 2025-11-27 09:36:29.279 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:29 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4544 DF PROTO=TCP SPT=35184 DPT=9101 SEQ=2397825623 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12963D1A0000000001030307) Nov 27 04:36:29 localhost systemd[1]: var-lib-containers-storage-overlay-f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048-merged.mount: Deactivated successfully. Nov 27 04:36:29 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:30 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:30 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:30 localhost systemd[1]: var-lib-containers-storage-overlay-3edfdc699753a1c833a1247909047263cd4d267465db29104ef571eb019dbe34-merged.mount: Deactivated successfully. Nov 27 04:36:30 localhost nova_compute[232618]: 2025-11-27 09:36:30.861 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:32 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=36355 DF PROTO=TCP SPT=42340 DPT=9105 SEQ=3673889381 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A1296495B0000000001030307) Nov 27 04:36:32 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:33 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:33 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:33 localhost systemd[1]: var-lib-containers-storage-overlay-f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048-merged.mount: Deactivated successfully. Nov 27 04:36:33 localhost systemd[1]: var-lib-containers-storage-overlay-f04f6aa8018da724c9daa5ca37db7cd13477323f1b725eec5dac97862d883048-merged.mount: Deactivated successfully. Nov 27 04:36:33 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:34 localhost nova_compute[232618]: 2025-11-27 09:36:34.322 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:34 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:36:34 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:34 localhost podman[243644]: 2025-11-27 09:36:34.904920953 +0000 UTC m=+0.079618033 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, container_name=multipathd, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, config_id=multipathd) Nov 27 04:36:34 localhost podman[243644]: 2025-11-27 09:36:34.94225478 +0000 UTC m=+0.116951810 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:36:34 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:35 localhost systemd[1]: var-lib-containers-storage-overlay-f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a-merged.mount: Deactivated successfully. Nov 27 04:36:35 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:35 localhost systemd[1]: var-lib-containers-storage-overlay-47afe78ba3ac18f156703d7ad9e4be64941a9d1bd472a4c2a59f4f2c3531ee35-merged.mount: Deactivated successfully. Nov 27 04:36:35 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:36:35 localhost nova_compute[232618]: 2025-11-27 09:36:35.865 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:36 localhost systemd[1]: var-lib-containers-storage-overlay-c892fd6b7d17c3244e97732d72b83cd3d1a569af20da04450edaf25f54095ce6-merged.mount: Deactivated successfully. Nov 27 04:36:36 localhost systemd[1]: var-lib-containers-storage-overlay-06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66-merged.mount: Deactivated successfully. Nov 27 04:36:36 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:36 localhost systemd[1]: var-lib-containers-storage-overlay-f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a-merged.mount: Deactivated successfully. Nov 27 04:36:37 localhost systemd[1]: var-lib-containers-storage-overlay-cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa-merged.mount: Deactivated successfully. Nov 27 04:36:37 localhost systemd[1]: var-lib-containers-storage-overlay-cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa-merged.mount: Deactivated successfully. Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.650 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.650 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.674 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.674 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.674 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.944 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.944 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.944 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:36:37 localhost nova_compute[232618]: 2025-11-27 09:36:37.944 232622 DEBUG nova.objects.instance [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:36:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4545 DF PROTO=TCP SPT=35184 DPT=9101 SEQ=2397825623 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12965DDB0000000001030307) Nov 27 04:36:38 localhost systemd[1]: var-lib-containers-storage-overlay-b574f97f279779c52df37c61d993141d596fdb6544fa700fbddd8f35f27a4d3b-merged.mount: Deactivated successfully. Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.731 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.751 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.751 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.751 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.752 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.752 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.752 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.753 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.753 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.753 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.754 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:36:38 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32922 DF PROTO=TCP SPT=47996 DPT=9882 SEQ=3036162168 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129660370000000001030307) Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.772 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.772 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.772 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.773 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:36:38 localhost nova_compute[232618]: 2025-11-27 09:36:38.773 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:36:38 localhost podman[243501]: 2025-11-27 09:36:38.816339957 +0000 UTC m=+14.874981230 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:36:38 localhost podman[243501]: 2025-11-27 09:36:38.845761881 +0000 UTC m=+14.904403124 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent) Nov 27 04:36:39 localhost systemd[1]: var-lib-containers-storage-overlay-3edfdc699753a1c833a1247909047263cd4d267465db29104ef571eb019dbe34-merged.mount: Deactivated successfully. Nov 27 04:36:39 localhost systemd[1]: var-lib-containers-storage-overlay-55e8351b1958513e36671035d0bb47863b4e87c80590aade3f4e58207a1d6315-merged.mount: Deactivated successfully. Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.246 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.473s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.312 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.313 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.360 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:39 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.515 232622 WARNING nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.516 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=12355MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.516 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.516 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.585 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.585 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.585 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:36:39 localhost nova_compute[232618]: 2025-11-27 09:36:39.642 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.069 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.427s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.076 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.093 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.094 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.095 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.579s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:36:40 localhost nova_compute[232618]: 2025-11-27 09:36:40.868 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:40 localhost systemd[1]: var-lib-containers-storage-overlay-d6636e8195e20b46e9ff0be91c525681b79b061d34e7042a3302554bc91c2a8c-merged.mount: Deactivated successfully. Nov 27 04:36:40 localhost systemd[1]: var-lib-containers-storage-overlay-cf8de856f68682579de884f5a9ccb4b00fffe375a72087325354c97a26c55ce7-merged.mount: Deactivated successfully. Nov 27 04:36:41 localhost systemd[1]: var-lib-containers-storage-overlay-cf8de856f68682579de884f5a9ccb4b00fffe375a72087325354c97a26c55ce7-merged.mount: Deactivated successfully. Nov 27 04:36:41 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=32924 DF PROTO=TCP SPT=47996 DPT=9882 SEQ=3036162168 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12966C5A0000000001030307) Nov 27 04:36:41 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:41 localhost systemd[1]: var-lib-containers-storage-overlay-3edfdc699753a1c833a1247909047263cd4d267465db29104ef571eb019dbe34-merged.mount: Deactivated successfully. Nov 27 04:36:42 localhost systemd[1]: var-lib-containers-storage-overlay-3edfdc699753a1c833a1247909047263cd4d267465db29104ef571eb019dbe34-merged.mount: Deactivated successfully. Nov 27 04:36:42 localhost systemd[1]: var-lib-containers-storage-overlay-d6636e8195e20b46e9ff0be91c525681b79b061d34e7042a3302554bc91c2a8c-merged.mount: Deactivated successfully. Nov 27 04:36:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:36:43.542 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:36:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:36:43.543 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:36:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:36:43.544 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:36:43 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:43 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:36:43 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:43 localhost podman[243735]: 2025-11-27 09:36:43.81676699 +0000 UTC m=+0.073558903 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=starting, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:36:43 localhost podman[243735]: 2025-11-27 09:36:43.821268505 +0000 UTC m=+0.078060448 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_id=edpm, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0) Nov 27 04:36:43 localhost podman[243735]: unhealthy Nov 27 04:36:43 localhost systemd[1]: var-lib-containers-storage-overlay-f49a20fc1f5020138578527318ecbf7083cb8c7be7c4014409c81f2cedb36958-merged.mount: Deactivated successfully. Nov 27 04:36:44 localhost kernel: overlayfs: lowerdir is in-use as upperdir/workdir of another mount, accessing files from both mounts will result in undefined behavior. Nov 27 04:36:44 localhost podman[243512]: 2025-11-27 09:36:24.383229435 +0000 UTC m=+0.050844246 image pull quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7 Nov 27 04:36:44 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Main process exited, code=exited, status=1/FAILURE Nov 27 04:36:44 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Failed with result 'exit-code'. Nov 27 04:36:44 localhost nova_compute[232618]: 2025-11-27 09:36:44.387 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:36:44 localhost systemd[1]: var-lib-containers-storage-overlay-c892fd6b7d17c3244e97732d72b83cd3d1a569af20da04450edaf25f54095ce6-merged.mount: Deactivated successfully. Nov 27 04:36:45 localhost systemd[1]: var-lib-containers-storage-overlay-efd486ab4cd4ff83f3804626a19ad34bc69aaee72db0852b1e52409f0ff23ebf-merged.mount: Deactivated successfully. Nov 27 04:36:45 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:45 localhost systemd[1]: var-lib-containers-storage-overlay-3df44265ee334241877fc90da4598858e128dcd022ea76b8f6ef87bd0d8667ae-merged.mount: Deactivated successfully. Nov 27 04:36:45 localhost kernel: overlayfs: lowerdir is in-use as upperdir/workdir of another mount, accessing files from both mounts will result in undefined behavior. Nov 27 04:44:10 localhost python3.9[266411]: ansible-ansible.legacy.dnf Invoked with name=['iscsi-initiator-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:44:10 localhost rsyslogd[760]: imjournal: 8185 messages lost due to rate-limiting (20000 allowed within 600 seconds) Nov 27 04:44:11 localhost nova_compute[232618]: 2025-11-27 09:44:11.037 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:12 localhost nova_compute[232618]: 2025-11-27 09:44:12.662 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:44:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:44:14 localhost systemd[1]: tmp-crun.uxJ8Ib.mount: Deactivated successfully. Nov 27 04:44:14 localhost podman[266525]: 2025-11-27 09:44:14.698014536 +0000 UTC m=+0.092769804 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:44:14 localhost podman[266525]: 2025-11-27 09:44:14.735689006 +0000 UTC m=+0.130444254 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:44:14 localhost systemd[1]: tmp-crun.q7cJGx.mount: Deactivated successfully. Nov 27 04:44:14 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:44:14 localhost podman[266524]: 2025-11-27 09:44:14.750066934 +0000 UTC m=+0.144860803 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, container_name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:44:14 localhost podman[266524]: 2025-11-27 09:44:14.809892546 +0000 UTC m=+0.204686375 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:44:14 localhost python3.9[266523]: ansible-ansible.builtin.stat Invoked with path=/var/lib/config-data/puppet-generated/iscsid/etc/iscsi follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:44:14 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:44:15 localhost ovn_controller[156436]: 2025-11-27T09:44:15Z|00055|memory_trim|INFO|Detected inactivity (last active 30001 ms ago): trimming memory Nov 27 04:44:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47484 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D58990000000001030307) Nov 27 04:44:15 localhost python3.9[266680]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/restorecon -nvr /etc/iscsi /var/lib/iscsi _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:44:16 localhost nova_compute[232618]: 2025-11-27 09:44:16.041 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47485 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D5C9B0000000001030307) Nov 27 04:44:16 localhost python3.9[266791]: ansible-ansible.builtin.stat Invoked with path=/etc/iscsi/.initiator_reset follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:44:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45456 DF PROTO=TCP SPT=50036 DPT=9102 SEQ=243315685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D5FDA0000000001030307) Nov 27 04:44:17 localhost nova_compute[232618]: 2025-11-27 09:44:17.664 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:17 localhost python3.9[266903]: ansible-ansible.builtin.lineinfile Invoked with insertafter=^#node.session.auth.chap.algs line=node.session.auth.chap_algs = SHA3-256,SHA256,SHA1,MD5 path=/etc/iscsi/iscsid.conf regexp=^node.session.auth.chap_algs state=present encoding=utf-8 backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47486 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D649A0000000001030307) Nov 27 04:44:19 localhost python3.9[267013]: ansible-ansible.builtin.systemd_service Invoked with enabled=True name=iscsid.socket state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:44:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=4641 DF PROTO=TCP SPT=48174 DPT=9102 SEQ=2764150335 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D67DB0000000001030307) Nov 27 04:44:19 localhost python3.9[267125]: ansible-ansible.builtin.systemd_service Invoked with enabled=True name=iscsid state=started daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:44:21 localhost nova_compute[232618]: 2025-11-27 09:44:21.045 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:22 localhost python3.9[267237]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:44:22 localhost network[267254]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:44:22 localhost network[267255]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:44:22 localhost network[267256]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:44:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47487 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D745A0000000001030307) Nov 27 04:44:22 localhost nova_compute[232618]: 2025-11-27 09:44:22.666 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:24 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:44:25 localhost openstack_network_exporter[244641]: ERROR 09:44:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:44:25 localhost openstack_network_exporter[244641]: ERROR 09:44:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:44:25 localhost openstack_network_exporter[244641]: Nov 27 04:44:25 localhost openstack_network_exporter[244641]: ERROR 09:44:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:44:25 localhost openstack_network_exporter[244641]: ERROR 09:44:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:44:25 localhost openstack_network_exporter[244641]: ERROR 09:44:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:44:25 localhost openstack_network_exporter[244641]: Nov 27 04:44:26 localhost nova_compute[232618]: 2025-11-27 09:44:26.046 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:27 localhost nova_compute[232618]: 2025-11-27 09:44:27.668 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:27 localhost python3.9[267490]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/modules-load.d selevel=s0 setype=etc_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 04:44:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:44:28 localhost podman[267601]: 2025-11-27 09:44:28.621772447 +0000 UTC m=+0.066867828 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, container_name=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:44:28 localhost podman[267601]: 2025-11-27 09:44:28.660108953 +0000 UTC m=+0.105204364 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 04:44:28 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:44:28 localhost python3.9[267600]: ansible-community.general.modprobe Invoked with name=dm-multipath state=present params= persistent=disabled Nov 27 04:44:29 localhost python3.9[267728]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/dm-multipath.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:44:30 localhost podman[267785]: 2025-11-27 09:44:30.346791922 +0000 UTC m=+0.085157594 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:44:30 localhost podman[267785]: 2025-11-27 09:44:30.381912298 +0000 UTC m=+0.120278000 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible) Nov 27 04:44:30 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:44:30 localhost python3.9[267786]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/etc/modules-load.d/dm-multipath.conf _original_basename=module-load.conf.j2 recurse=False state=file path=/etc/modules-load.d/dm-multipath.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:30 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47488 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129D93DA0000000001030307) Nov 27 04:44:31 localhost nova_compute[232618]: 2025-11-27 09:44:31.065 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:31 localhost python3.9[267913]: ansible-ansible.builtin.lineinfile Invoked with create=True dest=/etc/modules line=dm-multipath mode=0644 state=present path=/etc/modules encoding=utf-8 backrefs=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:32 localhost python3.9[268023]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/multipath setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:32 localhost nova_compute[232618]: 2025-11-27 09:44:32.670 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:44:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:44:32 localhost systemd[1]: tmp-crun.CzHEh4.mount: Deactivated successfully. Nov 27 04:44:32 localhost podman[268135]: 2025-11-27 09:44:32.962776894 +0000 UTC m=+0.098514917 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-type=git, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, version=9.6, config_id=edpm, io.openshift.expose-services=, release=1755695350, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, managed_by=edpm_ansible, maintainer=Red Hat, Inc., name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., io.buildah.version=1.33.7, architecture=x86_64, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:44:32 localhost podman[268135]: 2025-11-27 09:44:32.999308035 +0000 UTC m=+0.135046048 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, maintainer=Red Hat, Inc., version=9.6, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, managed_by=edpm_ansible, io.buildah.version=1.33.7, com.redhat.component=ubi9-minimal-container, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, vendor=Red Hat, Inc., architecture=x86_64, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, release=1755695350, build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:44:33 localhost podman[268134]: 2025-11-27 09:44:33.013195211 +0000 UTC m=+0.151747184 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 04:44:33 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:44:33 localhost podman[268134]: 2025-11-27 09:44:33.030858912 +0000 UTC m=+0.169410915 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 04:44:33 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:44:33 localhost python3.9[268133]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:44:34 localhost python3.9[268284]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:44:34 localhost python3.9[268396]: ansible-ansible.legacy.command Invoked with _raw_params=grep -q '^blacklist\s*{' /etc/multipath.conf _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:44:35 localhost python3.9[268507]: ansible-ansible.builtin.replace Invoked with path=/etc/multipath.conf regexp=^blacklist\s*{\n[\s]+devnode \"\.\*\" replace=blacklist { backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:35 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.7 (252 of 333 items), suggesting rotation. Nov 27 04:44:35 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:44:35 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:44:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:44:35 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:44:35 localhost podman[268509]: 2025-11-27 09:44:35.990709518 +0000 UTC m=+0.087065242 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:44:35 localhost podman[268509]: 2025-11-27 09:44:35.996688937 +0000 UTC m=+0.093044691 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:44:36 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:44:36 localhost nova_compute[232618]: 2025-11-27 09:44:36.105 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:36 localhost python3.9[268641]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= find_multipaths yes path=/etc/multipath.conf regexp=^\s+find_multipaths state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:37 localhost python3.9[268751]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= recheck_wwid yes path=/etc/multipath.conf regexp=^\s+recheck_wwid state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:37 localhost nova_compute[232618]: 2025-11-27 09:44:37.672 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:37 localhost python3.9[268861]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= skip_kpartx yes path=/etc/multipath.conf regexp=^\s+skip_kpartx state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.454 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.455 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.455 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.456 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.456 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:44:38 localhost podman[242678]: time="2025-11-27T09:44:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:44:38 localhost podman[242678]: @ - - [27/Nov/2025:09:44:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 149986 "" "Go-http-client/1.1" Nov 27 04:44:38 localhost podman[242678]: @ - - [27/Nov/2025:09:44:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 17704 "" "Go-http-client/1.1" Nov 27 04:44:38 localhost python3.9[268991]: ansible-ansible.builtin.lineinfile Invoked with firstmatch=True insertafter=^defaults line= user_friendly_names no path=/etc/multipath.conf regexp=^\s+user_friendly_names state=present encoding=utf-8 backrefs=False create=False backup=False unsafe_writes=False search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:38 localhost nova_compute[232618]: 2025-11-27 09:44:38.916 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.460s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.159 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.160 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.375 232622 WARNING nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.377 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11834MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.378 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.378 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.512 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.512 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.513 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:44:39 localhost nova_compute[232618]: 2025-11-27 09:44:39.555 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:44:40 localhost nova_compute[232618]: 2025-11-27 09:44:40.017 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.462s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:44:40 localhost nova_compute[232618]: 2025-11-27 09:44:40.025 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:44:40 localhost nova_compute[232618]: 2025-11-27 09:44:40.064 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:44:40 localhost nova_compute[232618]: 2025-11-27 09:44:40.066 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:44:40 localhost nova_compute[232618]: 2025-11-27 09:44:40.067 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.689s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:44:40 localhost python3.9[269125]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath.conf follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:44:41 localhost nova_compute[232618]: 2025-11-27 09:44:41.136 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:41 localhost python3.9[269237]: ansible-ansible.builtin.file Invoked with path=/var/local/libexec recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:42 localhost nova_compute[232618]: 2025-11-27 09:44:42.674 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:42 localhost python3.9[269347]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-container-shutdown follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:43 localhost nova_compute[232618]: 2025-11-27 09:44:43.068 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:43 localhost nova_compute[232618]: 2025-11-27 09:44:43.068 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:43 localhost nova_compute[232618]: 2025-11-27 09:44:43.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:43 localhost nova_compute[232618]: 2025-11-27 09:44:43.261 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:43 localhost python3.9[269404]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-container-shutdown _original_basename=edpm-container-shutdown recurse=False state=file path=/var/local/libexec/edpm-container-shutdown force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.549 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.549 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.550 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.550 162092 ERROR neutron.agent.linux.external_process [-] metadata-proxy for metadata with uuid 3d77aad9-8375-4ff0-8227-0972f0af9ab9 not found. The process should not have died#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.550 162092 WARNING neutron.agent.linux.external_process [-] Respawning metadata-proxy for uuid 3d77aad9-8375-4ff0-8227-0972f0af9ab9#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.550 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.551 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[039c03d4-222a-4c98-9a46-ee4644fb52ac]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.552 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: global Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9 Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: user root Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: group root Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: daemon Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: defaults Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: log global Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: mode http Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: option httplog Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: listen listener Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID 3d77aad9-8375-4ff0-8227-0972f0af9ab9 Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 04:44:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:43.553 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'env', 'PROCESS_TAG=haproxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/3d77aad9-8375-4ff0-8227-0972f0af9ab9.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 04:44:43 localhost podman[269518]: Nov 27 04:44:43 localhost podman[269518]: 2025-11-27 09:44:43.976303507 +0000 UTC m=+0.092722004 container create 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:44:44 localhost systemd[1]: Started libpod-conmon-0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014.scope. Nov 27 04:44:44 localhost podman[269518]: 2025-11-27 09:44:43.934578036 +0000 UTC m=+0.050996553 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 04:44:44 localhost systemd[1]: Started libcrun container. Nov 27 04:44:44 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/56bac76312ede99d3db3a0fd8d60c45c239702a13f7e02223b811f27ac89ef11/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 04:44:44 localhost podman[269518]: 2025-11-27 09:44:44.05544563 +0000 UTC m=+0.171864117 container init 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:44:44 localhost podman[269518]: 2025-11-27 09:44:44.06549691 +0000 UTC m=+0.181915397 container start 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 04:44:44 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [NOTICE] (269558) : New worker (269560) forked Nov 27 04:44:44 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [NOTICE] (269558) : Loading success. Nov 27 04:44:44 localhost ovn_metadata_agent[162087]: 2025-11-27 09:44:44.121 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.571s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:44:44 localhost python3.9[269549]: ansible-ansible.legacy.stat Invoked with path=/var/local/libexec/edpm-start-podman-container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:44 localhost nova_compute[232618]: 2025-11-27 09:44:44.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:44 localhost nova_compute[232618]: 2025-11-27 09:44:44.260 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:44:44 localhost python3.9[269625]: ansible-ansible.legacy.file Invoked with group=root mode=0700 owner=root setype=container_file_t dest=/var/local/libexec/edpm-start-podman-container _original_basename=edpm-start-podman-container recurse=False state=file path=/var/local/libexec/edpm-start-podman-container force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:44:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:44:44 localhost podman[269659]: 2025-11-27 09:44:44.993415569 +0000 UTC m=+0.085854222 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 04:44:45 localhost podman[269659]: 2025-11-27 09:44:45.032888134 +0000 UTC m=+0.125326767 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:44:45 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:44:45 localhost podman[269660]: 2025-11-27 09:44:45.045270302 +0000 UTC m=+0.137502670 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:44:45 localhost podman[269660]: 2025-11-27 09:44:45.084913 +0000 UTC m=+0.177145348 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:44:45 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:44:45 localhost nova_compute[232618]: 2025-11-27 09:44:45.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:45 localhost nova_compute[232618]: 2025-11-27 09:44:45.261 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 04:44:45 localhost nova_compute[232618]: 2025-11-27 09:44:45.290 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 04:44:45 localhost python3.9[269783]: ansible-ansible.builtin.file Invoked with mode=420 path=/etc/systemd/system-preset state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11976 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DCDC80000000001030307) Nov 27 04:44:46 localhost python3.9[269893]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/edpm-container-shutdown.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:46 localhost nova_compute[232618]: 2025-11-27 09:44:46.175 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:46 localhost nova_compute[232618]: 2025-11-27 09:44:46.285 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:46 localhost nova_compute[232618]: 2025-11-27 09:44:46.285 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11977 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DD1DB0000000001030307) Nov 27 04:44:46 localhost python3.9[269950]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/edpm-container-shutdown.service _original_basename=edpm-container-shutdown-service recurse=False state=file path=/etc/systemd/system/edpm-container-shutdown.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47489 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DD3DA0000000001030307) Nov 27 04:44:47 localhost nova_compute[232618]: 2025-11-27 09:44:47.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:47 localhost nova_compute[232618]: 2025-11-27 09:44:47.260 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 04:44:47 localhost nova_compute[232618]: 2025-11-27 09:44:47.274 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:47 localhost python3.9[270060]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:47 localhost nova_compute[232618]: 2025-11-27 09:44:47.676 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:47 localhost python3.9[270117]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-edpm-container-shutdown.preset _original_basename=91-edpm-container-shutdown-preset recurse=False state=file path=/etc/systemd/system-preset/91-edpm-container-shutdown.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:48 localhost nova_compute[232618]: 2025-11-27 09:44:48.287 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:48 localhost nova_compute[232618]: 2025-11-27 09:44:48.287 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:44:48 localhost nova_compute[232618]: 2025-11-27 09:44:48.288 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:44:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11978 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DD9DA0000000001030307) Nov 27 04:44:49 localhost python3.9[270228]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=edpm-container-shutdown state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:44:49 localhost systemd[1]: Reloading. Nov 27 04:44:49 localhost systemd-rc-local-generator[270252]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:44:49 localhost systemd-sysv-generator[270258]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:49 localhost nova_compute[232618]: 2025-11-27 09:44:49.418 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:44:49 localhost nova_compute[232618]: 2025-11-27 09:44:49.419 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:44:49 localhost nova_compute[232618]: 2025-11-27 09:44:49.420 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:44:49 localhost nova_compute[232618]: 2025-11-27 09:44:49.420 232622 DEBUG nova.objects.instance [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:44:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45457 DF PROTO=TCP SPT=50036 DPT=9102 SEQ=243315685 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DDDDA0000000001030307) Nov 27 04:44:50 localhost nova_compute[232618]: 2025-11-27 09:44:50.052 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:44:50 localhost nova_compute[232618]: 2025-11-27 09:44:50.081 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:44:50 localhost nova_compute[232618]: 2025-11-27 09:44:50.081 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:44:50 localhost python3.9[270376]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system/netns-placeholder.service follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:50 localhost python3.9[270433]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system/netns-placeholder.service _original_basename=netns-placeholder-service recurse=False state=file path=/etc/systemd/system/netns-placeholder.service force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:51 localhost nova_compute[232618]: 2025-11-27 09:44:51.051 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:44:51 localhost nova_compute[232618]: 2025-11-27 09:44:51.202 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:51 localhost python3.9[270543]: ansible-ansible.legacy.stat Invoked with path=/etc/systemd/system-preset/91-netns-placeholder.preset follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:51 localhost python3.9[270600]: ansible-ansible.legacy.file Invoked with group=root mode=0644 owner=root dest=/etc/systemd/system-preset/91-netns-placeholder.preset _original_basename=91-netns-placeholder-preset recurse=False state=file path=/etc/systemd/system-preset/91-netns-placeholder.preset force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:44:52 localhost nova_compute[232618]: 2025-11-27 09:44:52.677 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11979 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129DE99A0000000001030307) Nov 27 04:44:53 localhost python3.9[270710]: ansible-ansible.builtin.systemd Invoked with daemon_reload=True enabled=True name=netns-placeholder state=started daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:44:53 localhost systemd[1]: Reloading. Nov 27 04:44:53 localhost systemd-sysv-generator[270737]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:44:53 localhost systemd-rc-local-generator[270733]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:44:53 localhost systemd[1]: Starting Create netns directory... Nov 27 04:44:53 localhost systemd[1]: run-netns-placeholder.mount: Deactivated successfully. Nov 27 04:44:53 localhost systemd[1]: netns-placeholder.service: Deactivated successfully. Nov 27 04:44:53 localhost systemd[1]: Finished Create netns directory. Nov 27 04:44:54 localhost python3.9[270862]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/healthchecks setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:55 localhost openstack_network_exporter[244641]: ERROR 09:44:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:44:55 localhost openstack_network_exporter[244641]: ERROR 09:44:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:44:55 localhost openstack_network_exporter[244641]: ERROR 09:44:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:44:55 localhost openstack_network_exporter[244641]: ERROR 09:44:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:44:55 localhost openstack_network_exporter[244641]: Nov 27 04:44:55 localhost openstack_network_exporter[244641]: ERROR 09:44:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:44:55 localhost openstack_network_exporter[244641]: Nov 27 04:44:56 localhost python3.9[271021]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/healthchecks/multipathd/healthcheck follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:56 localhost nova_compute[232618]: 2025-11-27 09:44:56.240 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:57 localhost nova_compute[232618]: 2025-11-27 09:44:57.678 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:44:57 localhost python3.9[271115]: ansible-ansible.legacy.file Invoked with group=zuul mode=0700 owner=zuul setype=container_file_t dest=/var/lib/openstack/healthchecks/multipathd/ _original_basename=healthcheck recurse=False state=file path=/var/lib/openstack/healthchecks/multipathd/ force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:58 localhost python3.9[271225]: ansible-ansible.builtin.file Invoked with path=/var/lib/kolla/config_files recurse=True setype=container_file_t state=directory force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:44:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:44:59 localhost podman[271243]: 2025-11-27 09:44:59.004701781 +0000 UTC m=+0.089942074 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:44:59 localhost podman[271243]: 2025-11-27 09:44:59.043989421 +0000 UTC m=+0.129229654 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=multipathd, io.buildah.version=1.41.3) Nov 27 04:44:59 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:44:59 localhost python3.9[271356]: ansible-ansible.legacy.stat Invoked with path=/var/lib/kolla/config_files/multipathd.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:44:59 localhost python3.9[271413]: ansible-ansible.legacy.file Invoked with mode=0600 dest=/var/lib/kolla/config_files/multipathd.json _original_basename=.7u8qkm0t recurse=False state=file path=/var/lib/kolla/config_files/multipathd.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:45:00 localhost systemd[1]: tmp-crun.CnGPJ8.mount: Deactivated successfully. Nov 27 04:45:00 localhost podman[271524]: 2025-11-27 09:45:00.673560956 +0000 UTC m=+0.087239937 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:45:00 localhost podman[271524]: 2025-11-27 09:45:00.70699781 +0000 UTC m=+0.120676871 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, managed_by=edpm_ansible) Nov 27 04:45:00 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:45:00 localhost python3.9[271523]: ansible-ansible.builtin.file Invoked with mode=0755 path=/var/lib/edpm-config/container-startup-config/multipathd state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11980 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E09DB0000000001030307) Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.281 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.843 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.862 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.862 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.862 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:45:01 localhost nova_compute[232618]: 2025-11-27 09:45:01.909 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.046s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:45:02 localhost nova_compute[232618]: 2025-11-27 09:45:02.680 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:03 localhost python3.9[271817]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/edpm-config/container-startup-config/multipathd config_pattern=*.json debug=False Nov 27 04:45:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:45:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:45:03 localhost systemd[1]: tmp-crun.WPOGR6.mount: Deactivated successfully. Nov 27 04:45:03 localhost podman[271928]: 2025-11-27 09:45:03.903837716 +0000 UTC m=+0.097148604 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 04:45:03 localhost podman[271928]: 2025-11-27 09:45:03.939167457 +0000 UTC m=+0.132478355 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, tcib_managed=true, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm) Nov 27 04:45:03 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:45:03 localhost systemd[1]: tmp-crun.1bnHza.mount: Deactivated successfully. Nov 27 04:45:03 localhost podman[271929]: 2025-11-27 09:45:03.996724132 +0000 UTC m=+0.187151858 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, config_id=edpm, maintainer=Red Hat, Inc., name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, io.openshift.expose-services=, version=9.6, io.buildah.version=1.33.7, vcs-type=git, io.openshift.tags=minimal rhel9, distribution-scope=public, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, managed_by=edpm_ansible, architecture=x86_64) Nov 27 04:45:04 localhost podman[271929]: 2025-11-27 09:45:04.01391224 +0000 UTC m=+0.204339936 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, io.buildah.version=1.33.7, managed_by=edpm_ansible, release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, build-date=2025-08-20T13:12:41, name=ubi9-minimal, version=9.6, io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., config_id=edpm, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:45:04 localhost python3.9[271927]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:45:04 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:45:05 localhost python3.9[272075]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Nov 27 04:45:06 localhost nova_compute[232618]: 2025-11-27 09:45:06.325 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:45:06 localhost podman[272120]: 2025-11-27 09:45:06.991565719 +0000 UTC m=+0.088189949 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:45:07 localhost podman[272120]: 2025-11-27 09:45:07.004049271 +0000 UTC m=+0.100673481 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:45:07 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:45:07 localhost nova_compute[232618]: 2025-11-27 09:45:07.682 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:08 localhost podman[242678]: time="2025-11-27T09:45:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:45:08 localhost podman[242678]: @ - - [27/Nov/2025:09:45:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:45:08 localhost podman[242678]: @ - - [27/Nov/2025:09:45:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18180 "" "Go-http-client/1.1" Nov 27 04:45:10 localhost python3[272235]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/edpm-config/container-startup-config/multipathd config_id=multipathd config_overrides={} config_patterns=*.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:45:10 localhost python3[272235]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "f275b8d168f7f57f31e3da49224019f39f95c80a833f083696a964527b07b54f",#012 "Digest": "sha256:6296d2d95faaeb90443ee98443b39aa81b5152414f9542335d72711bb15fefdd",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-multipathd:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6296d2d95faaeb90443ee98443b39aa81b5152414f9542335d72711bb15fefdd"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:12:42.268223466Z",#012 "Config": {#012 "User": "root",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 249482220,#012 "VirtualSize": 249482220,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/da9f726a106a4f4af24ed404443eca5cd50a43c6e5c864c256f158761c28e938/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/da9f726a106a4f4af24ed404443eca5cd50a43c6e5c864c256f158761c28e938/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:135e1f5eea0bd6ac73fc43c122f58d5ed97cb8a56365c4a958c72d470055986b"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "root",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:37.752912815Z",#012 "created_by": "/bin/sh -c dnf install -y ca-certificates dumb-init glibc-langpack-en procps-ng python3 sudo util-linux-user which python-tcib-containers",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:38.066850603Z",#012 Nov 27 04:45:11 localhost nova_compute[232618]: 2025-11-27 09:45:11.352 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:11 localhost python3.9[272410]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:45:12 localhost python3.9[272522]: ansible-file Invoked with path=/etc/systemd/system/edpm_multipathd.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:12 localhost nova_compute[232618]: 2025-11-27 09:45:12.708 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:12 localhost python3.9[272577]: ansible-stat Invoked with path=/etc/systemd/system/edpm_multipathd_healthcheck.timer follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:45:13 localhost python3.9[272686]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236712.8638186-1367-273595683661954/source dest=/etc/systemd/system/edpm_multipathd.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:14 localhost python3.9[272741]: ansible-systemd Invoked with state=started name=edpm_multipathd.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51326 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E42F90000000001030307) Nov 27 04:45:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:45:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:45:16 localhost podman[272853]: 2025-11-27 09:45:15.997278054 +0000 UTC m=+0.085633987 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:45:16 localhost podman[272853]: 2025-11-27 09:45:16.031821705 +0000 UTC m=+0.120177648 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:45:16 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:45:16 localhost podman[272852]: 2025-11-27 09:45:16.050647955 +0000 UTC m=+0.142814912 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller) Nov 27 04:45:16 localhost podman[272852]: 2025-11-27 09:45:16.089812381 +0000 UTC m=+0.181979288 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:45:16 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:45:16 localhost python3.9[272851]: ansible-ansible.builtin.stat Invoked with path=/etc/multipath/.multipath_restart_required follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:45:16 localhost nova_compute[232618]: 2025-11-27 09:45:16.354 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51327 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E471A0000000001030307) Nov 27 04:45:16 localhost python3.9[273006]: ansible-ansible.builtin.file Invoked with path=/etc/multipath/.multipath_restart_required state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11981 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E49DA0000000001030307) Nov 27 04:45:17 localhost nova_compute[232618]: 2025-11-27 09:45:17.727 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:18 localhost python3.9[273116]: ansible-ansible.builtin.file Invoked with mode=0755 path=/etc/modules-load.d selevel=s0 setype=etc_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None attributes=None Nov 27 04:45:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51328 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E4F1A0000000001030307) Nov 27 04:45:18 localhost python3.9[273226]: ansible-community.general.modprobe Invoked with name=nvme-fabrics state=present params= persistent=disabled Nov 27 04:45:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=47490 DF PROTO=TCP SPT=37376 DPT=9102 SEQ=1558873262 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E51DA0000000001030307) Nov 27 04:45:19 localhost python3.9[273336]: ansible-ansible.legacy.stat Invoked with path=/etc/modules-load.d/nvme-fabrics.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:45:20 localhost python3.9[273393]: ansible-ansible.legacy.file Invoked with mode=0644 dest=/etc/modules-load.d/nvme-fabrics.conf _original_basename=module-load.conf.j2 recurse=False state=file path=/etc/modules-load.d/nvme-fabrics.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:21 localhost nova_compute[232618]: 2025-11-27 09:45:21.399 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:21 localhost python3.9[273503]: ansible-ansible.builtin.lineinfile Invoked with create=True dest=/etc/modules line=nvme-fabrics mode=0644 state=present path=/etc/modules encoding=utf-8 backrefs=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:22 localhost python3.9[273613]: ansible-ansible.legacy.dnf Invoked with name=['nvme-cli'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Nov 27 04:45:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51329 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E5EDB0000000001030307) Nov 27 04:45:22 localhost nova_compute[232618]: 2025-11-27 09:45:22.764 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:25 localhost openstack_network_exporter[244641]: ERROR 09:45:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:45:25 localhost openstack_network_exporter[244641]: ERROR 09:45:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:45:25 localhost openstack_network_exporter[244641]: ERROR 09:45:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:45:25 localhost openstack_network_exporter[244641]: ERROR 09:45:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:45:25 localhost openstack_network_exporter[244641]: Nov 27 04:45:25 localhost openstack_network_exporter[244641]: ERROR 09:45:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:45:25 localhost openstack_network_exporter[244641]: Nov 27 04:45:26 localhost python3.9[273723]: ansible-ansible.builtin.setup Invoked with gather_subset=['!all', '!min', 'local'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Nov 27 04:45:26 localhost nova_compute[232618]: 2025-11-27 09:45:26.402 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:27 localhost python3.9[273837]: ansible-ansible.builtin.file Invoked with mode=0644 path=/etc/ssh/ssh_known_hosts state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:27 localhost nova_compute[232618]: 2025-11-27 09:45:27.773 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:28 localhost python3.9[273947]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:45:28 localhost systemd[1]: Reloading. Nov 27 04:45:28 localhost systemd-rc-local-generator[273972]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:45:28 localhost systemd-sysv-generator[273978]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:28 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:29 localhost python3.9[274091]: ansible-ansible.builtin.service_facts Invoked Nov 27 04:45:29 localhost network[274108]: You are using 'network' service provided by 'network-scripts', which are now deprecated. Nov 27 04:45:29 localhost network[274109]: 'network-scripts' will be removed from distribution in near future. Nov 27 04:45:29 localhost network[274110]: It is advised to switch to 'NetworkManager' instead for network management. Nov 27 04:45:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:45:29 localhost podman[274116]: 2025-11-27 09:45:29.834147669 +0000 UTC m=+0.081571255 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:45:29 localhost podman[274116]: 2025-11-27 09:45:29.846822685 +0000 UTC m=+0.094246271 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_id=multipathd, io.buildah.version=1.41.3, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:45:30 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:45:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:45:30 localhost podman[274155]: 2025-11-27 09:45:30.851317983 +0000 UTC m=+0.081646367 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, managed_by=edpm_ansible, tcib_managed=true) Nov 27 04:45:30 localhost podman[274155]: 2025-11-27 09:45:30.8868666 +0000 UTC m=+0.117194974 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:45:30 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:45:31 localhost systemd[1]: /usr/lib/systemd/system/insights-client.service:23: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:45:31 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51330 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129E7FDB0000000001030307) Nov 27 04:45:31 localhost nova_compute[232618]: 2025-11-27 09:45:31.434 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:32 localhost nova_compute[232618]: 2025-11-27 09:45:32.796 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:45:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:45:34 localhost podman[274379]: 2025-11-27 09:45:34.609426885 +0000 UTC m=+0.084852737 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 04:45:34 localhost podman[274379]: 2025-11-27 09:45:34.61965478 +0000 UTC m=+0.095080672 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:45:34 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:45:34 localhost podman[274380]: 2025-11-27 09:45:34.707230184 +0000 UTC m=+0.176289287 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, container_name=openstack_network_exporter, version=9.6, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, io.openshift.expose-services=, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, build-date=2025-08-20T13:12:41, config_id=edpm, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, architecture=x86_64, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:45:34 localhost podman[274380]: 2025-11-27 09:45:34.721889399 +0000 UTC m=+0.190948502 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, release=1755695350, build-date=2025-08-20T13:12:41, vcs-type=git, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, io.buildah.version=1.33.7, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, distribution-scope=public, managed_by=edpm_ansible, name=ubi9-minimal) Nov 27 04:45:34 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:45:34 localhost python3.9[274378]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_compute.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:35 localhost python3.9[274528]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_migration_target.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:36 localhost nova_compute[232618]: 2025-11-27 09:45:36.436 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:36 localhost python3.9[274639]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_api_cron.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:37 localhost python3.9[274750]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_api.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:45:37 localhost systemd[1]: tmp-crun.THBSGs.mount: Deactivated successfully. Nov 27 04:45:37 localhost podman[274752]: 2025-11-27 09:45:37.409893777 +0000 UTC m=+0.071188286 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:45:37 localhost podman[274752]: 2025-11-27 09:45:37.417886236 +0000 UTC m=+0.079180685 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:45:37 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:45:37 localhost nova_compute[232618]: 2025-11-27 09:45:37.799 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:38 localhost python3.9[274882]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_conductor.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:38 localhost podman[242678]: time="2025-11-27T09:45:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:45:38 localhost podman[242678]: @ - - [27/Nov/2025:09:45:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:45:38 localhost podman[242678]: @ - - [27/Nov/2025:09:45:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18192 "" "Go-http-client/1.1" Nov 27 04:45:39 localhost python3.9[274993]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_metadata.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:40 localhost python3.9[275104]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_scheduler.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.286 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.286 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.287 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.287 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.288 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.732 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.444s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.794 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:45:40 localhost nova_compute[232618]: 2025-11-27 09:45:40.794 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:45:40 localhost python3.9[275235]: ansible-ansible.builtin.systemd_service Invoked with enabled=False name=tripleo_nova_vnc_proxy.service state=stopped daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.022 232622 WARNING nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.023 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11795MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.024 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.025 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.215 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.216 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.216 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.286 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.361 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.362 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.392 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.415 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: HW_CPU_X86_ABM,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_AVX2,COMPUTE_TRUSTED_CERTS,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_AMD_SVM,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_RESCUE_BFV,COMPUTE_VOLUME_ATTACH_WITH_TAG,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_AMI,HW_CPU_X86_AVX,HW_CPU_X86_BMI,COMPUTE_STORAGE_BUS_FDC,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_VIOMMU_MODEL_AUTO,COMPUTE_STORAGE_BUS_SATA,COMPUTE_NET_VIF_MODEL_E1000E,COMPUTE_IMAGE_TYPE_ARI,HW_CPU_X86_SVM,COMPUTE_NET_VIF_MODEL_VMXNET3,COMPUTE_GRAPHICS_MODEL_BOCHS,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_NET_VIF_MODEL_NE2K_PCI,HW_CPU_X86_BMI2,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_NET_VIF_MODEL_PCNET,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_SECURITY_TPM_2_0,HW_CPU_X86_FMA3,HW_CPU_X86_SSE2,HW_CPU_X86_SSE,HW_CPU_X86_AESNI,COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_GRAPHICS_MODEL_CIRRUS,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_DEVICE_TAGGING,COMPUTE_VOLUME_EXTEND,HW_CPU_X86_SSE4A,COMPUTE_STORAGE_BUS_IDE,HW_CPU_X86_MMX,HW_CPU_X86_SSE41,COMPUTE_ACCELERATORS,HW_CPU_X86_SSE42,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,HW_CPU_X86_SSSE3,HW_CPU_X86_F16C,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_VIOMMU_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_VGA,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_RTL8139,HW_CPU_X86_CLMUL,COMPUTE_STORAGE_BUS_USB _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.456 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.470 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.937 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.481s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.945 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.967 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.969 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:45:41 localhost nova_compute[232618]: 2025-11-27 09:45:41.970 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.945s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:45:42 localhost python3.9[275370]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:42 localhost nova_compute[232618]: 2025-11-27 09:45:42.843 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:43 localhost python3.9[275480]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:45:43.551 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:45:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:45:43.551 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:45:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:45:43.552 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:45:43 localhost python3.9[275590]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_api_cron.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:43 localhost nova_compute[232618]: 2025-11-27 09:45:43.970 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:43 localhost nova_compute[232618]: 2025-11-27 09:45:43.970 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:43 localhost nova_compute[232618]: 2025-11-27 09:45:43.971 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:43 localhost nova_compute[232618]: 2025-11-27 09:45:43.971 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.131 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.132 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.145 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.146 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '781bcb1c-69d1-4045-96ac-91342bd3cd99', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.132333', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7d9a2e6-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '8c5e0325c37c6bd38123954036fe32b2589f8b0f5e4094c04c72dc7122e3edbd'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.132333', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7d9bf74-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '1784c0378b4b9c261c0e0decad9591cbdca772419379103969619db8bd578930'}]}, 'timestamp': '2025-11-27 09:45:44.147162', '_unique_id': '3cc52582ccdc43ae90ab1ccca62f7605'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.150 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.186 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 974916304 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.186 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 177343418 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6a57726b-bafe-44e2-bd78-9507288d4463', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 974916304, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.150235', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7dfc3ce-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '20414b39f27b2d6d584042a711d0b08b2f9a3f66705fa896079ac4e05aecb9fc'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 177343418, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.150235', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7dfd8b4-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '567a0209f63639ebb6c437ff14208ebcb5566bbd4be6fb30b4907cd52b3e37aa'}]}, 'timestamp': '2025-11-27 09:45:44.187112', '_unique_id': '3f1a36091e4c4e849cb14eef4de4a11e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.189 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.189 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.193 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 130 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '05baef98-690b-4f74-979e-02bde1c65eb2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 130, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.189816', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e0d660-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': '164ef0e9639f9c8a8e961e11d96b10e6b2dfcf5c51539e68f2fc1d24a2280452'}]}, 'timestamp': '2025-11-27 09:45:44.193675', '_unique_id': '4763ae90b9a042ecb6436203c3d1f53d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.195 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.195 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c4e83fe0-2f35-4725-bffa-02729e19b6d3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.195947', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e14366-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': 'b01a166488f681f0d36a2a8e353e66436c1fa5856341944a92cab337b7a4faaf'}]}, 'timestamp': '2025-11-27 09:45:44.196418', '_unique_id': '8c5f647cd2514a7896e8af2edb9a58b1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.198 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.215 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 52.46875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8cf152f2-fe02-481b-bc73-2e2622918958', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 52.46875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:45:44.198646', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'd7e452b8-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.40392677, 'message_signature': 'b099aeb07b120ec3d5523431a3f25d527fb4be8a5b53aea7b2c5624c7f94978a'}]}, 'timestamp': '2025-11-27 09:45:44.216458', '_unique_id': '6d0d6d3b585843b6adc6b11a0bbce3e8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.218 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.218 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 9355 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'cad4eecb-b5e3-42bd-8bb9-15a03c68163c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9355, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.218712', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e4bcb2-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': 'c32390cfad9d203e7d63bd92efe22811b3470bf308a3ecf34c56e7ad1ed8040d'}]}, 'timestamp': '2025-11-27 09:45:44.219183', '_unique_id': '6a9d5712b248427e93cc5c69d0a5510b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.220 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.221 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.221 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 530 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd5372aef-4149-4c7d-b1da-9f99959b8201', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 530, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.221329', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e523f0-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': 'cc13a651d142a7cc7914f9a0df544d606f7732ff9936237b796acaf070655470'}]}, 'timestamp': '2025-11-27 09:45:44.221826', '_unique_id': '0bf0b845715d4a2f9448bb2df749d499'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.223 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.224 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.224 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 73908224 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.224 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '92e16314-fc5e-4832-95fd-6a36c7ea5eae', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 73908224, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.224132', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7e59060-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '188a1716f46f4a092cef7b51f9406a6e7c6529d852f7f40625d1b0bbd9a679c1'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.224132', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7e5a258-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '4d246b7201d56dfe57d118042f9ea3a08fcbc416a174ccba10e74907593bae0b'}]}, 'timestamp': '2025-11-27 09:45:44.225033', '_unique_id': '91e6a7caa06c4b52b2be596fd941c672'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.227 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1064 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 222 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '09957352-b9e1-4de4-b965-283185b2fb94', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1064, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.227419', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7e61288-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '58225bd2ea824f3fb20738fd75c668df20aeeeb31e1d4a4e7103529ad01c5f88'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 222, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.227419', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7e62368-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': 'aa0b7f37111b63c4e00307fbcb341d13e857ae09e61c418417b2c984d2f216d3'}]}, 'timestamp': '2025-11-27 09:45:44.228366', '_unique_id': 'a7d86d98d0a64255b51e9f686fe3b157'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.230 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eed8f066-27b7-4182-b512-3b52df04a74b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.230685', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e69046-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': 'b5ad868b00a73f6aaba53edeb1d34e3717e720c7a7f58fdae2f0c95899ed0298'}]}, 'timestamp': '2025-11-27 09:45:44.231150', '_unique_id': 'ae5728eff4c84f13bf000b144d7bf602'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '20f01a3b-d79b-419d-baa2-10b3833e498a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.233325', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e6f856-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': '879c3b97e523c007d1a6a83a7bbc0462778596be6801cfd4bdfc4accc72721e4'}]}, 'timestamp': '2025-11-27 09:45:44.233817', '_unique_id': 'f1fa20e19c7f40dd870c329a618c592e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.235 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.236 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd58443df-c2cb-480c-812d-d0e409c444cb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.236255', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7e769ee-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '996e4c4955425e606a2d42d0d0f7e31704aa61fd5265f464f3c927fae118ec5e'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.236255', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7e77bd2-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '71e92e624ef0f64815a9e0c1d75437aae965ab6dd615207534d829cee120f918'}]}, 'timestamp': '2025-11-27 09:45:44.237154', '_unique_id': 'd454ae5affb0461f8b1a3c3cf29c86ce'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.239 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 53310000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd3d6e427-2f33-49e5-af1d-66f438e26519', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 53310000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:45:44.239337', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'd7e7e310-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.40392677, 'message_signature': '25159fb8c9fc5bad5fafcd1b814bb727180556e8ec7bcd5d071c63fcecc0d669'}]}, 'timestamp': '2025-11-27 09:45:44.239812', '_unique_id': '10c8f8f89086436eaee2bc3fc5da76fd'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.241 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 29130240 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 4300800 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '67f67e3d-3295-4937-a189-b426a469b064', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 29130240, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.241959', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7e84878-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '707474091b450dd1234cc1cdf89cfe994368ef289db38608d7a34b1f2439cfef'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 4300800, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.241959', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7e85a16-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '227e2a5b356dc73e313a6ff92bb3dc51159640b68af7ad6f189fad7e06fdb5cb'}]}, 'timestamp': '2025-11-27 09:45:44.242842', '_unique_id': 'd2c996fb417a4b99ab1aa1890bfe6750'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.244 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 11314 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7aca8ac3-8fd0-4e8c-a432-555a5a8ae2d2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 11314, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.244992', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e8bf10-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': '39b41f70b404c54e56d6ad7947f999fddff57c207f1c8a0dc88bc21164141f4e'}]}, 'timestamp': '2025-11-27 09:45:44.245452', '_unique_id': 'a927108293b14525b7b5c0d5d3cd0a8b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 90 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e77db2ce-fa12-4e47-a3ae-8c0f84fdc686', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 90, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.247624', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7e925d6-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': '891063fe7826cee7120589c5eb0952c88809403ad6d74344a036d1bb6ebfe9fb'}]}, 'timestamp': '2025-11-27 09:45:44.248084', '_unique_id': '2fcf9898f07c432a8b6235b9a0cbc710'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 203265086 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 25604091 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e1e14dbf-e249-4f9a-a155-57bf1fa96bca', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 203265086, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.250200', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7e98a26-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': 'f863690b9cf00c6a68c74ba70b1a89e0bcfdc2ed3bb791f973d299bfe477b089'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 25604091, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.250200', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7e99c0a-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': '1802d522e59761512f2e64c9698de1eb0acf899dc8e7162f078dd0843190b6ac'}]}, 'timestamp': '2025-11-27 09:45:44.251083', '_unique_id': 'abd662313c014cc191996d9f0315e332'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e0a35778-4b3a-4989-b2fe-119cd21cfd14', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.253231', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7ea00be-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': 'ddec72394b8aff8cc224bc26f4242caca57635f171c9320e11635e0aaf45f2e9'}]}, 'timestamp': '2025-11-27 09:45:44.253720', '_unique_id': '112788aa5c734cd2bf6b5c95d9012026'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8a355d3e-1bfd-48f5-9513-08097a775676', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:45:44.255458', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'd7ea5564-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.377892026, 'message_signature': '45615cf698f0d631ff362949e520b3684c3e736a4dfddecccfe4e84569ce00e5'}]}, 'timestamp': '2025-11-27 09:45:44.255770', '_unique_id': 'c66a2e8000344abb873d2396c004028a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 498 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '93babfce-92a2-454a-957b-7e400c0010fd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 498, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.257061', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7ea92ea-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': 'e7ebcd43f1e1cd1ccdffd975cc614d7621f06a71253bdd5af97d5eeaf1254dc7'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.257061', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7ea9d08-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.338304411, 'message_signature': 'b1a9b801928dd5e6f33788c17f8026ce5a6fefadccdddae096ef553299e60c03'}]}, 'timestamp': '2025-11-27 09:45:44.257606', '_unique_id': '202b3510cce8451f9ef3cfd122cca337'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f8e0f0ed-128e-4487-a2e7-54c3aa31314d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:45:44.258949', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'd7eadc8c-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '87896b1c37850380af5fffb8ec86743825939590af8c05ef260c281515b4f7e1'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:45:44.258949', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'd7eae678-cb75-11f0-8166-fa163e6bfee1', 'monotonic_time': 10974.320392053, 'message_signature': '336ce86a4d0ddebab768dd66d7ce966e2bbcd0d147c6c68fcf172909aab9f71a'}]}, 'timestamp': '2025-11-27 09:45:44.259464', '_unique_id': '85956ca2f19044d6bcd4b5259325974b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:45:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:45:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 04:45:44 localhost python3.9[275700]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_api.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:44 localhost python3.9[275810]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_conductor.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:45 localhost python3.9[275920]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_metadata.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23142 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EB8290000000001030307) Nov 27 04:45:46 localhost python3.9[276030]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_scheduler.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:46 localhost nova_compute[232618]: 2025-11-27 09:45:46.257 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:46 localhost nova_compute[232618]: 2025-11-27 09:45:46.259 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:46 localhost nova_compute[232618]: 2025-11-27 09:45:46.260 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:45:46 localhost nova_compute[232618]: 2025-11-27 09:45:46.459 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23143 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EBC1B0000000001030307) Nov 27 04:45:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:45:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:45:46 localhost podman[276142]: 2025-11-27 09:45:46.679102714 +0000 UTC m=+0.066144693 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:45:46 localhost podman[276142]: 2025-11-27 09:45:46.686513492 +0000 UTC m=+0.073555501 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:45:46 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:45:46 localhost podman[276141]: 2025-11-27 09:45:46.744091137 +0000 UTC m=+0.130281554 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:45:46 localhost python3.9[276140]: ansible-ansible.builtin.file Invoked with path=/usr/lib/systemd/system/tripleo_nova_vnc_proxy.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:46 localhost podman[276141]: 2025-11-27 09:45:46.80838777 +0000 UTC m=+0.194578127 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:45:46 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:45:47 localhost nova_compute[232618]: 2025-11-27 09:45:47.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51331 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EBFDA0000000001030307) Nov 27 04:45:47 localhost python3.9[276297]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_compute.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:47 localhost nova_compute[232618]: 2025-11-27 09:45:47.846 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:48 localhost python3.9[276407]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_migration_target.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23144 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EC41A0000000001030307) Nov 27 04:45:48 localhost python3.9[276517]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_api_cron.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=11982 DF PROTO=TCP SPT=57182 DPT=9102 SEQ=3091083283 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EC7DA0000000001030307) Nov 27 04:45:50 localhost nova_compute[232618]: 2025-11-27 09:45:50.261 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:45:50 localhost nova_compute[232618]: 2025-11-27 09:45:50.261 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:45:50 localhost nova_compute[232618]: 2025-11-27 09:45:50.261 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:45:50 localhost python3.9[276627]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_api.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:50 localhost python3.9[276737]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_conductor.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:51 localhost nova_compute[232618]: 2025-11-27 09:45:51.294 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:45:51 localhost nova_compute[232618]: 2025-11-27 09:45:51.295 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:45:51 localhost nova_compute[232618]: 2025-11-27 09:45:51.295 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:45:51 localhost nova_compute[232618]: 2025-11-27 09:45:51.295 232622 DEBUG nova.objects.instance [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:45:51 localhost nova_compute[232618]: 2025-11-27 09:45:51.462 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:51 localhost python3.9[276847]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_metadata.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:52 localhost nova_compute[232618]: 2025-11-27 09:45:52.419 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:45:52 localhost nova_compute[232618]: 2025-11-27 09:45:52.435 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:45:52 localhost nova_compute[232618]: 2025-11-27 09:45:52.435 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:45:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23145 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129ED3DA0000000001030307) Nov 27 04:45:52 localhost nova_compute[232618]: 2025-11-27 09:45:52.882 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:52 localhost python3.9[276957]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_scheduler.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:53 localhost python3.9[277067]: ansible-ansible.builtin.file Invoked with path=/etc/systemd/system/tripleo_nova_vnc_proxy.service state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:45:54 localhost python3.9[277177]: ansible-ansible.legacy.command Invoked with _raw_params=if systemctl is-active certmonger.service; then#012 systemctl disable --now certmonger.service#012 test -f /etc/systemd/system/certmonger.service || systemctl mask certmonger.service#012fi#012 _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True cmd=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:45:55 localhost python3.9[277287]: ansible-ansible.builtin.find Invoked with file_type=any hidden=True paths=['/var/lib/certmonger/requests'] patterns=[] read_whole_file=False age_stamp=mtime recurse=False follow=False get_checksum=False checksum_algorithm=sha1 use_regex=False exact_mode=True excludes=None contains=None age=None size=None depth=None mode=None encoding=None limit=None Nov 27 04:45:55 localhost openstack_network_exporter[244641]: ERROR 09:45:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:45:55 localhost openstack_network_exporter[244641]: ERROR 09:45:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:45:55 localhost openstack_network_exporter[244641]: ERROR 09:45:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:45:55 localhost openstack_network_exporter[244641]: ERROR 09:45:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:45:55 localhost openstack_network_exporter[244641]: Nov 27 04:45:55 localhost openstack_network_exporter[244641]: ERROR 09:45:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:45:55 localhost openstack_network_exporter[244641]: Nov 27 04:45:56 localhost nova_compute[232618]: 2025-11-27 09:45:56.464 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:56 localhost python3.9[277397]: ansible-ansible.builtin.systemd_service Invoked with daemon_reload=True daemon_reexec=False scope=system no_block=False name=None state=None enabled=None force=None masked=None Nov 27 04:45:56 localhost systemd[1]: Reloading. Nov 27 04:45:56 localhost systemd-sysv-generator[277422]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:45:56 localhost systemd-rc-local-generator[277418]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:56 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:45:57 localhost python3.9[277578]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_compute.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:45:57 localhost nova_compute[232618]: 2025-11-27 09:45:57.885 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:45:58 localhost python3.9[277721]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_migration_target.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:45:59 localhost python3.9[277850]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_api_cron.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:00 localhost python3.9[277961]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_api.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:46:00 localhost podman[277963]: 2025-11-27 09:46:00.54131912 +0000 UTC m=+0.070239173 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:46:00 localhost podman[277963]: 2025-11-27 09:46:00.552188089 +0000 UTC m=+0.081108202 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_id=multipathd) Nov 27 04:46:00 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:46:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23146 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129EF3DA0000000001030307) Nov 27 04:46:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:46:01 localhost systemd[1]: tmp-crun.5JKY7J.mount: Deactivated successfully. Nov 27 04:46:01 localhost podman[278091]: 2025-11-27 09:46:01.024943198 +0000 UTC m=+0.090901543 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 04:46:01 localhost podman[278091]: 2025-11-27 09:46:01.058966666 +0000 UTC m=+0.124925011 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:46:01 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:46:01 localhost python3.9[278090]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_conductor.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:01 localhost nova_compute[232618]: 2025-11-27 09:46:01.517 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:02 localhost python3.9[278219]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_metadata.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:02 localhost nova_compute[232618]: 2025-11-27 09:46:02.917 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:03 localhost python3.9[278330]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_scheduler.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:46:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:46:04 localhost podman[278333]: 2025-11-27 09:46:04.990959808 +0000 UTC m=+0.083825295 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, managed_by=edpm_ansible, architecture=x86_64, release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.buildah.version=1.33.7, version=9.6, build-date=2025-08-20T13:12:41, maintainer=Red Hat, Inc., vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:46:05 localhost podman[278333]: 2025-11-27 09:46:05.036182223 +0000 UTC m=+0.129047690 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., container_name=openstack_network_exporter, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, config_id=edpm, maintainer=Red Hat, Inc., build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64) Nov 27 04:46:05 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:46:05 localhost podman[278332]: 2025-11-27 09:46:05.041813844 +0000 UTC m=+0.137559987 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 04:46:05 localhost podman[278332]: 2025-11-27 09:46:05.120764537 +0000 UTC m=+0.216510650 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:46:05 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:46:05 localhost python3.9[278480]: ansible-ansible.legacy.command Invoked with cmd=/usr/bin/systemctl reset-failed tripleo_nova_vnc_proxy.service _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True _raw_params=None argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:46:06 localhost nova_compute[232618]: 2025-11-27 09:46:06.518 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:46:07 localhost nova_compute[232618]: 2025-11-27 09:46:07.921 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:07 localhost podman[278553]: 2025-11-27 09:46:07.981788257 +0000 UTC m=+0.077130747 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:46:08 localhost podman[278553]: 2025-11-27 09:46:08.020935971 +0000 UTC m=+0.116278481 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:46:08 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:46:08 localhost podman[242678]: time="2025-11-27T09:46:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:46:08 localhost podman[242678]: @ - - [27/Nov/2025:09:46:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:46:08 localhost python3.9[278615]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:08 localhost podman[242678]: @ - - [27/Nov/2025:09:46:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18200 "" "Go-http-client/1.1" Nov 27 04:46:09 localhost python3.9[278725]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/containers setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:10 localhost python3.9[278835]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/openstack/config/nova_nvme_cleaner setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:11 localhost python3.9[278945]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/nova setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:11 localhost nova_compute[232618]: 2025-11-27 09:46:11.520 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:11 localhost python3.9[279055]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/_nova_secontext setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:12 localhost python3.9[279165]: ansible-ansible.builtin.file Invoked with group=zuul mode=0755 owner=zuul path=/var/lib/nova/instances setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:12 localhost nova_compute[232618]: 2025-11-27 09:46:12.957 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:13 localhost python3.9[279275]: ansible-ansible.builtin.file Invoked with group=root mode=0750 owner=root path=/etc/ceph setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:13 localhost python3.9[279385]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/etc/multipath setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:14 localhost python3.9[279495]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/etc/nvme setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:14 localhost python3.9[279605]: ansible-ansible.builtin.file Invoked with group=zuul owner=zuul path=/run/openvswitch setype=container_file_t state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39926 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F2D590000000001030307) Nov 27 04:46:16 localhost nova_compute[232618]: 2025-11-27 09:46:16.523 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39927 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F315B0000000001030307) Nov 27 04:46:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:46:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:46:16 localhost podman[279623]: 2025-11-27 09:46:16.9923802 +0000 UTC m=+0.080612059 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 04:46:17 localhost podman[279623]: 2025-11-27 09:46:17.037135563 +0000 UTC m=+0.125367432 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:46:17 localhost podman[279624]: 2025-11-27 09:46:17.05388424 +0000 UTC m=+0.139084178 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:46:17 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:46:17 localhost podman[279624]: 2025-11-27 09:46:17.068036507 +0000 UTC m=+0.153236435 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:46:17 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:46:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23147 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F33DA0000000001030307) Nov 27 04:46:17 localhost nova_compute[232618]: 2025-11-27 09:46:17.959 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39928 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F395A0000000001030307) Nov 27 04:46:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=51332 DF PROTO=TCP SPT=40756 DPT=9102 SEQ=2207632511 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F3DDA0000000001030307) Nov 27 04:46:21 localhost nova_compute[232618]: 2025-11-27 09:46:21.526 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:22 localhost python3.9[279762]: ansible-ansible.builtin.getent Invoked with database=passwd key=nova fail_key=True service=None split=None Nov 27 04:46:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39929 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F491B0000000001030307) Nov 27 04:46:22 localhost nova_compute[232618]: 2025-11-27 09:46:22.995 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:23 localhost sshd[279781]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:46:23 localhost systemd-logind[761]: New session 61 of user zuul. Nov 27 04:46:23 localhost systemd[1]: Started Session 61 of User zuul. Nov 27 04:46:23 localhost systemd[1]: session-61.scope: Deactivated successfully. Nov 27 04:46:23 localhost systemd-logind[761]: Session 61 logged out. Waiting for processes to exit. Nov 27 04:46:23 localhost systemd-logind[761]: Removed session 61. Nov 27 04:46:24 localhost python3.9[279892]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/config.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:25 localhost python3.9[279978]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/config.json mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236784.3386593-3040-92280116025641/.source.json follow=False _original_basename=config.json.j2 checksum=b51012bfb0ca26296dcf3793a2f284446fb1395e backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:25 localhost openstack_network_exporter[244641]: ERROR 09:46:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:46:25 localhost openstack_network_exporter[244641]: ERROR 09:46:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:46:25 localhost openstack_network_exporter[244641]: ERROR 09:46:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:46:25 localhost openstack_network_exporter[244641]: ERROR 09:46:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:46:25 localhost openstack_network_exporter[244641]: Nov 27 04:46:25 localhost openstack_network_exporter[244641]: ERROR 09:46:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:46:25 localhost openstack_network_exporter[244641]: Nov 27 04:46:25 localhost python3.9[280086]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/nova-blank.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:26 localhost python3.9[280141]: ansible-ansible.legacy.file Invoked with mode=0644 setype=container_file_t dest=/var/lib/openstack/config/nova/nova-blank.conf _original_basename=nova-blank.conf recurse=False state=file path=/var/lib/openstack/config/nova/nova-blank.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:26 localhost nova_compute[232618]: 2025-11-27 09:46:26.529 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:27 localhost python3.9[280249]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/ssh-config follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:27 localhost python3.9[280335]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/ssh-config mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236786.61675-3040-114134676227656/.source follow=False _original_basename=ssh-config checksum=4297f735c41bdc1ff52d72e6f623a02242f37958 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:28 localhost nova_compute[232618]: 2025-11-27 09:46:28.000 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:28 localhost python3.9[280443]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/02-nova-host-specific.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:29 localhost python3.9[280529]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/02-nova-host-specific.conf mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236787.8529775-3040-84745562444445/.source.conf follow=False _original_basename=02-nova-host-specific.conf.j2 checksum=ed20ceeae546fb32715585d43ee52c52842c4309 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:29 localhost python3.9[280637]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/nova_statedir_ownership.py follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:30 localhost python3.9[280723]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/nova_statedir_ownership.py mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236789.273619-3040-99563782464453/.source.py follow=False _original_basename=nova_statedir_ownership.py checksum=c6c8a3cfefa5efd60ceb1408c4e977becedb71e2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:46:30 localhost python3.9[280831]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/nova/run-on-host follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:30 localhost podman[280832]: 2025-11-27 09:46:30.978662011 +0000 UTC m=+0.076046497 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_id=multipathd) Nov 27 04:46:30 localhost podman[280832]: 2025-11-27 09:46:30.989657915 +0000 UTC m=+0.087042401 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=multipathd, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:46:30 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:46:31 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39930 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129F69DA0000000001030307) Nov 27 04:46:31 localhost python3.9[280934]: ansible-ansible.legacy.copy Invoked with dest=/var/lib/openstack/config/nova/run-on-host mode=0644 setype=container_file_t src=/home/zuul/.ansible/tmp/ansible-tmp-1764236790.4163802-3040-71143679692292/.source follow=False _original_basename=run-on-host checksum=93aba8edc83d5878604a66d37fea2f12b60bdea2 backup=False force=True remote_src=False unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:31 localhost nova_compute[232618]: 2025-11-27 09:46:31.566 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:46:31 localhost podman[280952]: 2025-11-27 09:46:31.988676668 +0000 UTC m=+0.080474465 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_metadata_agent) Nov 27 04:46:32 localhost podman[280952]: 2025-11-27 09:46:32.021801971 +0000 UTC m=+0.113599778 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 04:46:32 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:46:32 localhost python3.9[281063]: ansible-ansible.builtin.file Invoked with group=nova mode=0700 owner=nova path=/home/nova/.ssh state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:46:33 localhost nova_compute[232618]: 2025-11-27 09:46:33.038 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:33 localhost python3.9[281173]: ansible-ansible.legacy.copy Invoked with dest=/home/nova/.ssh/authorized_keys group=nova mode=0600 owner=nova remote_src=True src=/var/lib/openstack/config/nova/ssh-publickey backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:46:34 localhost python3.9[281283]: ansible-ansible.builtin.stat Invoked with path=/var/lib/nova/compute_id follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:46:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:46:35 localhost systemd[1]: tmp-crun.rBFL72.mount: Deactivated successfully. Nov 27 04:46:35 localhost podman[281395]: 2025-11-27 09:46:35.909065551 +0000 UTC m=+0.080654271 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:46:35 localhost podman[281395]: 2025-11-27 09:46:35.922969921 +0000 UTC m=+0.094558611 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0) Nov 27 04:46:35 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:46:35 localhost systemd[1]: tmp-crun.ZZNnhe.mount: Deactivated successfully. Nov 27 04:46:35 localhost podman[281397]: 2025-11-27 09:46:35.980813762 +0000 UTC m=+0.147266375 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, release=1755695350, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, vendor=Red Hat, Inc., architecture=x86_64, maintainer=Red Hat, Inc., build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_id=edpm, io.openshift.expose-services=, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, version=9.6, io.buildah.version=1.33.7) Nov 27 04:46:35 localhost podman[281397]: 2025-11-27 09:46:35.994012644 +0000 UTC m=+0.160465217 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.buildah.version=1.33.7, vcs-type=git, config_id=edpm, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.expose-services=, name=ubi9-minimal, vendor=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., container_name=openstack_network_exporter, architecture=x86_64, maintainer=Red Hat, Inc., io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 04:46:36 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:46:36 localhost python3.9[281396]: ansible-ansible.builtin.file Invoked with group=nova mode=0400 owner=nova path=/var/lib/nova/compute_id state=file recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:46:36 localhost nova_compute[232618]: 2025-11-27 09:46:36.568 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:36 localhost python3.9[281542]: ansible-ansible.builtin.stat Invoked with path=/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:37 localhost python3.9[281652]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/containers/nova_compute.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:38 localhost nova_compute[232618]: 2025-11-27 09:46:38.043 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:38 localhost python3.9[281707]: ansible-ansible.legacy.file Invoked with mode=0644 setype=container_file_t dest=/var/lib/openstack/config/containers/nova_compute.json _original_basename=nova_compute.json.j2 recurse=False state=file path=/var/lib/openstack/config/containers/nova_compute.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:38 localhost python3.9[281815]: ansible-ansible.legacy.stat Invoked with path=/var/lib/openstack/config/containers/nova_compute_init.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True get_selinux_context=False Nov 27 04:46:38 localhost podman[242678]: time="2025-11-27T09:46:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:46:38 localhost podman[242678]: @ - - [27/Nov/2025:09:46:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:46:38 localhost podman[242678]: @ - - [27/Nov/2025:09:46:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18194 "" "Go-http-client/1.1" Nov 27 04:46:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:46:38 localhost podman[281848]: 2025-11-27 09:46:38.98091634 +0000 UTC m=+0.073230183 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:46:38 localhost podman[281848]: 2025-11-27 09:46:38.991985195 +0000 UTC m=+0.084299048 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:46:39 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:46:39 localhost python3.9[281882]: ansible-ansible.legacy.file Invoked with mode=0700 setype=container_file_t dest=/var/lib/openstack/config/containers/nova_compute_init.json _original_basename=nova_compute_init.json.j2 recurse=False state=file path=/var/lib/openstack/config/containers/nova_compute_init.json force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None attributes=None Nov 27 04:46:40 localhost python3.9[282003]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/containers config_pattern=nova_compute_init.json debug=False Nov 27 04:46:41 localhost python3.9[282113]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.297 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.297 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.298 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.298 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.298 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.610 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.837 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.538s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.908 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:46:41 localhost nova_compute[232618]: 2025-11-27 09:46:41.909 232622 DEBUG nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.069 232622 WARNING nova.virt.libvirt.driver [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.070 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11812MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.070 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.070 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.196 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.197 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.197 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.253 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:46:42 localhost python3[282245]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/containers config_id=edpm config_overrides={} config_patterns=nova_compute_init.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:46:42 localhost python3[282245]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "b65793e7266422f5b94c32d109b906c8ffd974cf2ddf0b6929e463e29e05864a",#012 "Digest": "sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:36:07.10279245Z",#012 "Config": {#012 "User": "nova",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 1211782527,#012 "VirtualSize": 1211782527,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/c3914bdda39f47c0c497a56396d11c84b489b87df2bfd019b00ddced1e1ae309/diff:/var/lib/containers/storage/overlay/f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:c136b33417f134a3b932677bcf7a2df089c29f20eca250129eafd2132d4708bb",#012 "sha256:7913bde445307e7f24767d9149b2e7f498930793ac9f073ccec69b608c009d31",#012 "sha256:084b2323a717fe711217b0ec21da61f4804f7a0d506adae935888421b80809cf"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "nova",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 {#012 Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.720 232622 DEBUG oslo_concurrency.processutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.467s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.727 232622 DEBUG nova.compute.provider_tree [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.748 232622 DEBUG nova.scheduler.client.report [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.750 232622 DEBUG nova.compute.resource_tracker [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:46:42 localhost nova_compute[232618]: 2025-11-27 09:46:42.751 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.680s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:46:43 localhost nova_compute[232618]: 2025-11-27 09:46:43.082 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:46:43.553 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:46:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:46:43.554 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:46:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:46:43.555 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:46:43 localhost nova_compute[232618]: 2025-11-27 09:46:43.752 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:43 localhost nova_compute[232618]: 2025-11-27 09:46:43.753 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:43 localhost nova_compute[232618]: 2025-11-27 09:46:43.753 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:44 localhost nova_compute[232618]: 2025-11-27 09:46:44.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:44 localhost python3.9[282439]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52472 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FA2890000000001030307) Nov 27 04:46:45 localhost python3.9[282551]: ansible-container_config_data Invoked with config_overrides={} config_path=/var/lib/openstack/config/containers config_pattern=nova_compute.json debug=False Nov 27 04:46:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52473 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FA69B0000000001030307) Nov 27 04:46:46 localhost nova_compute[232618]: 2025-11-27 09:46:46.656 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:46:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:46:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39931 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FA9DA0000000001030307) Nov 27 04:46:47 localhost podman[282663]: 2025-11-27 09:46:47.517621995 +0000 UTC m=+0.081183875 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:46:47 localhost podman[282663]: 2025-11-27 09:46:47.526878481 +0000 UTC m=+0.090440291 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:46:47 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:46:47 localhost systemd[1]: tmp-crun.drQc4t.mount: Deactivated successfully. Nov 27 04:46:47 localhost podman[282662]: 2025-11-27 09:46:47.627820381 +0000 UTC m=+0.198002438 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=ovn_controller, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller) Nov 27 04:46:47 localhost python3.9[282661]: ansible-container_config_hash Invoked with check_mode=False config_vol_prefix=/var/lib/config-data Nov 27 04:46:47 localhost podman[282662]: 2025-11-27 09:46:47.66342105 +0000 UTC m=+0.233603047 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 04:46:47 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:46:48 localhost nova_compute[232618]: 2025-11-27 09:46:48.084 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:48 localhost nova_compute[232618]: 2025-11-27 09:46:48.256 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:48 localhost nova_compute[232618]: 2025-11-27 09:46:48.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:48 localhost nova_compute[232618]: 2025-11-27 09:46:48.260 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:46:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52474 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FAE9A0000000001030307) Nov 27 04:46:48 localhost python3[282820]: ansible-edpm_container_manage Invoked with concurrency=1 config_dir=/var/lib/openstack/config/containers config_id=edpm config_overrides={} config_patterns=nova_compute.json log_base_path=/var/log/containers/stdouts debug=False Nov 27 04:46:48 localhost python3[282820]: ansible-edpm_container_manage PODMAN-CONTAINER-DEBUG: [#012 {#012 "Id": "b65793e7266422f5b94c32d109b906c8ffd974cf2ddf0b6929e463e29e05864a",#012 "Digest": "sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e",#012 "RepoTags": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified"#012 ],#012 "RepoDigests": [#012 "quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:647f1d5dc1b70ffa3e1832199619d57bfaeceac8823ff53ece64b8e42cc9688e"#012 ],#012 "Parent": "",#012 "Comment": "",#012 "Created": "2025-11-26T06:36:07.10279245Z",#012 "Config": {#012 "User": "nova",#012 "Env": [#012 "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",#012 "LANG=en_US.UTF-8",#012 "TZ=UTC",#012 "container=oci"#012 ],#012 "Entrypoint": [#012 "dumb-init",#012 "--single-child",#012 "--"#012 ],#012 "Cmd": [#012 "kolla_start"#012 ],#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "StopSignal": "SIGTERM"#012 },#012 "Version": "",#012 "Author": "",#012 "Architecture": "amd64",#012 "Os": "linux",#012 "Size": 1211782527,#012 "VirtualSize": 1211782527,#012 "GraphDriver": {#012 "Name": "overlay",#012 "Data": {#012 "LowerDir": "/var/lib/containers/storage/overlay/c3914bdda39f47c0c497a56396d11c84b489b87df2bfd019b00ddced1e1ae309/diff:/var/lib/containers/storage/overlay/f20c3ba929bbb53a84e323dddb8c0eaf3ca74b6729310e964e1fa9eee119e43a/diff:/var/lib/containers/storage/overlay/06a1fa74af6494e3f3865876d25e5a11b62fb12ede8164b96bce734f8d084c66/diff:/var/lib/containers/storage/overlay/cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa/diff",#012 "UpperDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/diff",#012 "WorkDir": "/var/lib/containers/storage/overlay/f7726cecd9e8969401979ecd2369f385c53efc762aea19175eca5dfbffa00449/work"#012 }#012 },#012 "RootFS": {#012 "Type": "layers",#012 "Layers": [#012 "sha256:cf752d9babba20815c6849e3dd587209dffdfbbc56c600ddbc26d05721943ffa",#012 "sha256:1e3477d3ea795ca64b46f28aa9428ba791c4250e0fd05e173a4b9c0fb0bdee23",#012 "sha256:c136b33417f134a3b932677bcf7a2df089c29f20eca250129eafd2132d4708bb",#012 "sha256:7913bde445307e7f24767d9149b2e7f498930793ac9f073ccec69b608c009d31",#012 "sha256:084b2323a717fe711217b0ec21da61f4804f7a0d506adae935888421b80809cf"#012 ]#012 },#012 "Labels": {#012 "io.buildah.version": "1.41.3",#012 "maintainer": "OpenStack Kubernetes Operator team",#012 "org.label-schema.build-date": "20251125",#012 "org.label-schema.license": "GPLv2",#012 "org.label-schema.name": "CentOS Stream 9 Base Image",#012 "org.label-schema.schema-version": "1.0",#012 "org.label-schema.vendor": "CentOS",#012 "tcib_build_tag": "1f5c0439f2433cb462b222a5bb23e629",#012 "tcib_managed": "true"#012 },#012 "Annotations": {},#012 "ManifestType": "application/vnd.docker.distribution.manifest.v2+json",#012 "User": "nova",#012 "History": [#012 {#012 "created": "2025-11-25T04:02:36.223494528Z",#012 "created_by": "/bin/sh -c #(nop) ADD file:cacf1a97b4abfca5db2db22f7ddbca8fd7daa5076a559639c109f09aaf55871d in / ",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:36.223562059Z",#012 "created_by": "/bin/sh -c #(nop) LABEL org.label-schema.schema-version=\"1.0\" org.label-schema.name=\"CentOS Stream 9 Base Image\" org.label-schema.vendor=\"CentOS\" org.label-schema.license=\"GPLv2\" org.label-schema.build-date=\"20251125\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-25T04:02:39.054452717Z",#012 "created_by": "/bin/sh -c #(nop) CMD [\"/bin/bash\"]"#012 },#012 {#012 "created": "2025-11-26T06:10:57.55004106Z",#012 "created_by": "/bin/sh -c #(nop) LABEL maintainer=\"OpenStack Kubernetes Operator team\"",#012 "comment": "FROM quay.io/centos/centos:stream9",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550061231Z",#012 "created_by": "/bin/sh -c #(nop) LABEL tcib_managed=true",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550071761Z",#012 "created_by": "/bin/sh -c #(nop) ENV LANG=\"en_US.UTF-8\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550082711Z",#012 "created_by": "/bin/sh -c #(nop) ENV TZ=\"UTC\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550094371Z",#012 "created_by": "/bin/sh -c #(nop) ENV container=\"oci\"",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.550104472Z",#012 "created_by": "/bin/sh -c #(nop) USER root",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:10:57.937139683Z",#012 "created_by": "/bin/sh -c if [ -f \"/etc/yum.repos.d/ubi.repo\" ]; then rm -f /etc/yum.repos.d/ubi.repo && dnf clean all && rm -rf /var/cache/dnf; fi",#012 "empty_layer": true#012 },#012 {#012 "created": "2025-11-26T06:11:33.845342269Z",#012 "created_by": "/bin/sh -c dnf install -y crudini && crudini --del /etc/dnf/dnf.conf main override_install_langs && crudini --set /etc/dnf/dnf.conf main clean_requirements_on_remove True && crudini --set /etc/dnf/dnf.conf main exactarch 1 && crudini --set /etc/dnf/dnf.conf main gpgcheck 1 && crudini --set /etc/dnf/dnf.conf main install_weak_deps False && if [ 'centos' == 'centos' ];then crudini --set /etc/dnf/dnf.conf main best False; fi && crudini --set /etc/dnf/dnf.conf main installonly_limit 0 && crudini --set /etc/dnf/dnf.conf main keepcache 0 && crudini --set /etc/dnf/dnf.conf main obsoletes 1 && crudini --set /etc/dnf/dnf.conf main plugins 1 && crudini --set /etc/dnf/dnf.conf main skip_missing_names_on_install False && crudini --set /etc/dnf/dnf.conf main tsflags nodocs",#012 "empty_layer": true#012 },#012 {#012 Nov 27 04:46:49 localhost nova_compute[232618]: 2025-11-27 09:46:49.260 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=23148 DF PROTO=TCP SPT=50116 DPT=9102 SEQ=2783461103 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FB1DA0000000001030307) Nov 27 04:46:50 localhost python3.9[282994]: ansible-ansible.builtin.stat Invoked with path=/etc/sysconfig/podman_drop_in follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:50 localhost nova_compute[232618]: 2025-11-27 09:46:50.261 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:50 localhost nova_compute[232618]: 2025-11-27 09:46:50.261 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:46:50 localhost nova_compute[232618]: 2025-11-27 09:46:50.262 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:46:51 localhost python3.9[283106]: ansible-file Invoked with path=/etc/systemd/system/edpm_nova_compute.requires state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:46:51 localhost nova_compute[232618]: 2025-11-27 09:46:51.372 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:46:51 localhost nova_compute[232618]: 2025-11-27 09:46:51.373 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:46:51 localhost nova_compute[232618]: 2025-11-27 09:46:51.373 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:46:51 localhost nova_compute[232618]: 2025-11-27 09:46:51.373 232622 DEBUG nova.objects.instance [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:46:51 localhost nova_compute[232618]: 2025-11-27 09:46:51.660 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:51 localhost python3.9[283215]: ansible-copy Invoked with src=/home/zuul/.ansible/tmp/ansible-tmp-1764236811.2141376-3717-170777101898907/source dest=/etc/systemd/system/edpm_nova_compute.service mode=0644 owner=root group=root backup=False force=True remote_src=False follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:46:52 localhost nova_compute[232618]: 2025-11-27 09:46:52.384 232622 DEBUG nova.network.neutron [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:46:52 localhost nova_compute[232618]: 2025-11-27 09:46:52.403 232622 DEBUG oslo_concurrency.lockutils [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:46:52 localhost nova_compute[232618]: 2025-11-27 09:46:52.404 232622 DEBUG nova.compute.manager [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:46:52 localhost python3.9[283270]: ansible-systemd Invoked with state=started name=edpm_nova_compute.service enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:46:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52475 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FBE5A0000000001030307) Nov 27 04:46:53 localhost nova_compute[232618]: 2025-11-27 09:46:53.131 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:54 localhost nova_compute[232618]: 2025-11-27 09:46:54.400 232622 DEBUG oslo_service.periodic_task [None req-8804d2cf-3953-4173-8b51-10003ef21829 - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:46:54 localhost python3.9[283380]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner_healthcheck.service follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:55 localhost openstack_network_exporter[244641]: ERROR 09:46:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:46:55 localhost openstack_network_exporter[244641]: ERROR 09:46:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:46:55 localhost openstack_network_exporter[244641]: ERROR 09:46:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:46:55 localhost openstack_network_exporter[244641]: ERROR 09:46:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:46:55 localhost openstack_network_exporter[244641]: Nov 27 04:46:55 localhost openstack_network_exporter[244641]: ERROR 09:46:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:46:55 localhost openstack_network_exporter[244641]: Nov 27 04:46:55 localhost python3.9[283488]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner.service follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:56 localhost nova_compute[232618]: 2025-11-27 09:46:56.662 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:56 localhost python3.9[283596]: ansible-ansible.builtin.stat Invoked with path=/etc/systemd/system/edpm_nova_nvme_cleaner.service.requires follow=False get_checksum=True get_mime=True get_attributes=True get_selinux_context=False checksum_algorithm=sha1 Nov 27 04:46:58 localhost python3.9[283706]: ansible-containers.podman.podman_container Invoked with name=nova_nvme_cleaner state=absent executable=podman detach=True debug=False force_restart=False force_delete=True generate_systemd={} image_strict=False recreate=False image=None annotation=None arch=None attach=None authfile=None blkio_weight=None blkio_weight_device=None cap_add=None cap_drop=None cgroup_conf=None cgroup_parent=None cgroupns=None cgroups=None chrootdirs=None cidfile=None cmd_args=None conmon_pidfile=None command=None cpu_period=None cpu_quota=None cpu_rt_period=None cpu_rt_runtime=None cpu_shares=None cpus=None cpuset_cpus=None cpuset_mems=None decryption_key=None delete_depend=None delete_time=None delete_volumes=None detach_keys=None device=None device_cgroup_rule=None device_read_bps=None device_read_iops=None device_write_bps=None device_write_iops=None dns=None dns_option=None dns_search=None entrypoint=None env=None env_file=None env_host=None env_merge=None etc_hosts=None expose=None gidmap=None gpus=None group_add=None group_entry=None healthcheck=None healthcheck_interval=None healthcheck_retries=None healthcheck_start_period=None health_startup_cmd=None health_startup_interval=None health_startup_retries=None health_startup_success=None health_startup_timeout=None healthcheck_timeout=None healthcheck_failure_action=None hooks_dir=None hostname=None hostuser=None http_proxy=None image_volume=None init=None init_ctr=None init_path=None interactive=None ip=None ip6=None ipc=None kernel_memory=None label=None label_file=None log_driver=None log_level=None log_opt=None mac_address=None memory=None memory_reservation=None memory_swap=None memory_swappiness=None mount=None network=None network_aliases=None no_healthcheck=None no_hosts=None oom_kill_disable=None oom_score_adj=None os=None passwd=None passwd_entry=None personality=None pid=None pid_file=None pids_limit=None platform=None pod=None pod_id_file=None preserve_fd=None preserve_fds=None privileged=None publish=None publish_all=None pull=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None rdt_class=None read_only=None read_only_tmpfs=None requires=None restart_policy=None restart_time=None retry=None retry_delay=None rm=None rmi=None rootfs=None seccomp_policy=None secrets=NOT_LOGGING_PARAMETER sdnotify=None security_opt=None shm_size=None shm_size_systemd=None sig_proxy=None stop_signal=None stop_timeout=None stop_time=None subgidname=None subuidname=None sysctl=None systemd=None timeout=None timezone=None tls_verify=None tmpfs=None tty=None uidmap=None ulimit=None umask=None unsetenv=None unsetenv_all=None user=None userns=None uts=None variant=None volume=None volumes_from=None workdir=None Nov 27 04:46:58 localhost nova_compute[232618]: 2025-11-27 09:46:58.133 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:46:58 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 102.7 (342 of 333 items), suggesting rotation. Nov 27 04:46:58 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:46:58 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:46:58 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:46:58 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:46:59 localhost python3.9[283876]: ansible-ansible.builtin.systemd Invoked with name=edpm_nova_compute.service state=restarted daemon_reload=False daemon_reexec=False scope=system no_block=False enabled=None force=None masked=None Nov 27 04:46:59 localhost systemd[1]: Stopping nova_compute container... Nov 27 04:46:59 localhost nova_compute[232618]: 2025-11-27 09:46:59.349 232622 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170#033[00m Nov 27 04:47:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52476 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A129FDDDB0000000001030307) Nov 27 04:47:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:47:01 localhost podman[283925]: 2025-11-27 09:47:01.487474389 +0000 UTC m=+0.079626323 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, container_name=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 04:47:01 localhost podman[283925]: 2025-11-27 09:47:01.501879024 +0000 UTC m=+0.094030938 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.schema-version=1.0) Nov 27 04:47:01 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:47:01 localhost nova_compute[232618]: 2025-11-27 09:47:01.712 232622 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:47:02 localhost nova_compute[232618]: 2025-11-27 09:47:02.905 232622 WARNING amqp [-] Received method (60, 30) during closing channel 1. This method will be ignored#033[00m Nov 27 04:47:02 localhost nova_compute[232618]: 2025-11-27 09:47:02.909 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:47:02 localhost nova_compute[232618]: 2025-11-27 09:47:02.910 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:47:02 localhost nova_compute[232618]: 2025-11-27 09:47:02.910 232622 DEBUG oslo_concurrency.lockutils [None req-61eede4c-f320-4398-9dbb-d3b8655fde47 - - - - - -] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:47:02 localhost podman[283944]: 2025-11-27 09:47:02.995568941 +0000 UTC m=+0.089258130 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:47:03 localhost podman[283944]: 2025-11-27 09:47:03.004938601 +0000 UTC m=+0.098627780 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, managed_by=edpm_ansible, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:47:03 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:47:03 localhost journal[205316]: End of file while reading data: Input/output error Nov 27 04:47:03 localhost systemd[1]: libpod-4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6.scope: Deactivated successfully. Nov 27 04:47:03 localhost systemd[1]: libpod-4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6.scope: Consumed 20.472s CPU time. Nov 27 04:47:03 localhost podman[283901]: 2025-11-27 09:47:03.333383814 +0000 UTC m=+4.104782829 container died 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, config_id=edpm, org.label-schema.license=GPLv2, container_name=nova_compute, tcib_managed=true, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:47:03 localhost systemd[1]: tmp-crun.3SMgAx.mount: Deactivated successfully. Nov 27 04:47:03 localhost podman[283901]: 2025-11-27 09:47:03.501675979 +0000 UTC m=+4.273074934 container cleanup 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, tcib_managed=true, managed_by=edpm_ansible, config_id=edpm, container_name=nova_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:47:03 localhost podman[283901]: nova_compute Nov 27 04:47:03 localhost podman[284010]: error opening file `/run/crun/4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6/status`: No such file or directory Nov 27 04:47:03 localhost podman[283998]: 2025-11-27 09:47:03.570186115 +0000 UTC m=+0.046732106 container cleanup 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, managed_by=edpm_ansible, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, container_name=nova_compute, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:47:03 localhost podman[283998]: nova_compute Nov 27 04:47:03 localhost systemd[1]: edpm_nova_compute.service: Deactivated successfully. Nov 27 04:47:03 localhost systemd[1]: Stopped nova_compute container. Nov 27 04:47:03 localhost systemd[1]: Starting nova_compute container... Nov 27 04:47:03 localhost systemd[1]: Started libcrun container. Nov 27 04:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/nvme supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/etc/multipath supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/libvirt supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:03 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/be23feb758f9a5f6d373dfef988ab0b9302eb7ceb045d02de40fbc1cc5166d6c/merged/var/lib/iscsi supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:03 localhost podman[284012]: 2025-11-27 09:47:03.697237031 +0000 UTC m=+0.100368226 container init 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}, org.label-schema.schema-version=1.0, config_id=edpm, container_name=nova_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:47:03 localhost podman[284012]: 2025-11-27 09:47:03.706350525 +0000 UTC m=+0.109481720 container start 4c69ca77aa0707fe8adc387867db1b8c15d9845c7184f52c0b7d21093047aec6 (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute, container_name=nova_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=edpm, io.buildah.version=1.41.3, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': True, 'user': 'nova', 'restart': 'always', 'command': 'kolla_start', 'net': 'host', 'pid': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/var/lib/openstack/config/nova:/var/lib/kolla/config_files:ro', '/var/lib/openstack/cacerts/nova/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/etc/localtime:/etc/localtime:ro', '/lib/modules:/lib/modules:ro', '/dev:/dev', '/var/lib/libvirt:/var/lib/libvirt', '/run/libvirt:/run/libvirt:shared', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/etc/iscsi:/etc/iscsi:ro', '/etc/nvme:/etc/nvme', '/var/lib/openstack/config/ceph:/var/lib/kolla/config_files/ceph:ro', '/etc/ssh/ssh_known_hosts:/etc/ssh/ssh_known_hosts:ro']}) Nov 27 04:47:03 localhost podman[284012]: nova_compute Nov 27 04:47:03 localhost nova_compute[284026]: + sudo -E kolla_set_configs Nov 27 04:47:03 localhost systemd[1]: Started nova_compute container. Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Validating config file Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Kolla config strategy set to: COPY_ALWAYS Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying service configuration files Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/01-nova.conf to /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/01-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/03-ceph-nova.conf to /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/03-ceph-nova.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/99-nova-compute-cells-workarounds.conf to /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/99-nova-compute-cells-workarounds.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/nova-blank.conf to /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/nova-blank.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/02-nova-host-specific.conf to /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/nova/nova.conf.d/02-nova-host-specific.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /etc/ceph Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Creating directory /etc/ceph Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/ceph Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.conf to /etc/ceph/ceph.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/ceph/ceph.client.openstack.keyring to /etc/ceph/ceph.client.openstack.keyring Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-privatekey to /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /var/lib/nova/.ssh/config Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/ssh-config to /var/lib/nova/.ssh/config Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Deleting /usr/sbin/iscsiadm Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Copying /var/lib/kolla/config_files/run-on-host to /usr/sbin/iscsiadm Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /usr/sbin/iscsiadm Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Writing out command to execute Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/ceph/ceph.conf Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /etc/ceph/ceph.client.openstack.keyring Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:47:03 localhost nova_compute[284026]: INFO:__main__:Setting permission for /var/lib/nova/.ssh/config Nov 27 04:47:03 localhost nova_compute[284026]: ++ cat /run_command Nov 27 04:47:03 localhost nova_compute[284026]: + CMD=nova-compute Nov 27 04:47:03 localhost nova_compute[284026]: + ARGS= Nov 27 04:47:03 localhost nova_compute[284026]: + sudo kolla_copy_cacerts Nov 27 04:47:03 localhost nova_compute[284026]: + [[ ! -n '' ]] Nov 27 04:47:03 localhost nova_compute[284026]: + . kolla_extend_start Nov 27 04:47:03 localhost nova_compute[284026]: + echo 'Running command: '\''nova-compute'\''' Nov 27 04:47:03 localhost nova_compute[284026]: Running command: 'nova-compute' Nov 27 04:47:03 localhost nova_compute[284026]: + umask 0022 Nov 27 04:47:03 localhost nova_compute[284026]: + exec nova-compute Nov 27 04:47:04 localhost python3.9[284148]: ansible-containers.podman.podman_container Invoked with name=nova_compute_init state=started executable=podman detach=True debug=False force_restart=False force_delete=True generate_systemd={} image_strict=False recreate=False image=None annotation=None arch=None attach=None authfile=None blkio_weight=None blkio_weight_device=None cap_add=None cap_drop=None cgroup_conf=None cgroup_parent=None cgroupns=None cgroups=None chrootdirs=None cidfile=None cmd_args=None conmon_pidfile=None command=None cpu_period=None cpu_quota=None cpu_rt_period=None cpu_rt_runtime=None cpu_shares=None cpus=None cpuset_cpus=None cpuset_mems=None decryption_key=None delete_depend=None delete_time=None delete_volumes=None detach_keys=None device=None device_cgroup_rule=None device_read_bps=None device_read_iops=None device_write_bps=None device_write_iops=None dns=None dns_option=None dns_search=None entrypoint=None env=None env_file=None env_host=None env_merge=None etc_hosts=None expose=None gidmap=None gpus=None group_add=None group_entry=None healthcheck=None healthcheck_interval=None healthcheck_retries=None healthcheck_start_period=None health_startup_cmd=None health_startup_interval=None health_startup_retries=None health_startup_success=None health_startup_timeout=None healthcheck_timeout=None healthcheck_failure_action=None hooks_dir=None hostname=None hostuser=None http_proxy=None image_volume=None init=None init_ctr=None init_path=None interactive=None ip=None ip6=None ipc=None kernel_memory=None label=None label_file=None log_driver=None log_level=None log_opt=None mac_address=None memory=None memory_reservation=None memory_swap=None memory_swappiness=None mount=None network=None network_aliases=None no_healthcheck=None no_hosts=None oom_kill_disable=None oom_score_adj=None os=None passwd=None passwd_entry=None personality=None pid=None pid_file=None pids_limit=None platform=None pod=None pod_id_file=None preserve_fd=None preserve_fds=None privileged=None publish=None publish_all=None pull=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None rdt_class=None read_only=None read_only_tmpfs=None requires=None restart_policy=None restart_time=None retry=None retry_delay=None rm=None rmi=None rootfs=None seccomp_policy=None secrets=NOT_LOGGING_PARAMETER sdnotify=None security_opt=None shm_size=None shm_size_systemd=None sig_proxy=None stop_signal=None stop_timeout=None stop_time=None subgidname=None subuidname=None sysctl=None systemd=None timeout=None timezone=None tls_verify=None tmpfs=None tty=None uidmap=None ulimit=None umask=None unsetenv=None unsetenv_all=None user=None userns=None uts=None variant=None volume=None volumes_from=None workdir=None Nov 27 04:47:05 localhost systemd[1]: Started libpod-conmon-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope. Nov 27 04:47:05 localhost systemd[1]: Started libcrun container. Nov 27 04:47:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/usr/sbin/nova_statedir_ownership.py supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/var/lib/_nova_secontext supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:05 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae/merged/var/lib/nova supports timestamps until 2038 (0x7fffffff) Nov 27 04:47:05 localhost podman[284173]: 2025-11-27 09:47:05.27684426 +0000 UTC m=+0.130495409 container init 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, container_name=nova_compute_init, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 04:47:05 localhost podman[284173]: 2025-11-27 09:47:05.289491627 +0000 UTC m=+0.143142776 container start 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, org.label-schema.vendor=CentOS, config_id=edpm, container_name=nova_compute_init, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 04:47:05 localhost python3.9[284148]: ansible-containers.podman.podman_container PODMAN-CONTAINER-DEBUG: podman start nova_compute_init Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Applying nova statedir ownership Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Target ownership for /var/lib/nova: 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 1000 gid: 1000 path: /var/lib/nova/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Changing ownership of /var/lib/nova from 1000:1000 to 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 1000 gid: 1000 path: /var/lib/nova/instances/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Changing ownership of /var/lib/nova/instances from 1000:1000 to 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201 already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201 to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 0 gid: 0 path: /var/lib/nova/instances/a02f7f2f-d2cf-4612-b18a-8be435257201/console.log Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/_base already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/_base to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/454c8fd44a4104b436504db7d917f75244a9e94a Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/_base/ephemeral_1_0706d66 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/instances/locks already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/instances/locks to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/nova-454c8fd44a4104b436504db7d917f75244a9e94a Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/instances/locks/nova-ephemeral_1_0706d66 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 0 gid: 0 path: /var/lib/nova/delay-nova-compute Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/.ssh already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.ssh to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/ssh-privatekey Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.ssh/config Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/.cache already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.cache to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/ Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Ownership of /var/lib/nova/.cache/python-entrypoints already 42436:42436 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Setting selinux context of /var/lib/nova/.cache/python-entrypoints to system_u:object_r:container_file_t:s0 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/b234715fc878456b41e32c4fbc669b417044dbe6c6684bbc9059e5c93396ffea Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Checking uid: 42436 gid: 42436 path: /var/lib/nova/.cache/python-entrypoints/469bc4441baff9216df986857f9ff45dbf25965a8d2f755a6449ac2645cb7191 Nov 27 04:47:05 localhost nova_compute_init[284193]: INFO:nova_statedir:Nova statedir ownership complete Nov 27 04:47:05 localhost systemd[1]: libpod-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope: Deactivated successfully. Nov 27 04:47:05 localhost podman[284194]: 2025-11-27 09:47:05.369022086 +0000 UTC m=+0.059602899 container died 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, container_name=nova_compute_init, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 04:47:05 localhost podman[284208]: 2025-11-27 09:47:05.497243344 +0000 UTC m=+0.122316571 container cleanup 67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac (image=quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified, name=nova_compute_init, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified', 'privileged': False, 'user': 'root', 'restart': 'never', 'command': 'bash -c $* -- eval python3 /sbin/nova_statedir_ownership.py | logger -t nova_compute_init', 'net': 'none', 'security_opt': ['label=disable'], 'detach': False, 'environment': {'NOVA_STATEDIR_OWNERSHIP_SKIP': '/var/lib/nova/compute_id', '__OS_DEBUG': False}, 'volumes': ['/dev/log:/dev/log', '/var/lib/nova:/var/lib/nova:shared', '/var/lib/_nova_secontext:/var/lib/_nova_secontext:shared,z', '/var/lib/openstack/config/nova/nova_statedir_ownership.py:/sbin/nova_statedir_ownership.py:z']}, config_id=edpm, container_name=nova_compute_init, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0) Nov 27 04:47:05 localhost systemd[1]: libpod-conmon-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac.scope: Deactivated successfully. Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.586 284030 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.587 284030 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.587 284030 DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' initialize /usr/lib/python3.9/site-packages/os_vif/__init__.py:44#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.587 284030 INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.701 284030 DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.723 284030 DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 1 in 0.021s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:47:05 localhost nova_compute[284026]: 2025-11-27 09:47:05.723 284030 DEBUG oslo_concurrency.processutils [-] 'grep -F node.session.scan /sbin/iscsiadm' failed. Not Retrying. execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:473#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.141 284030 INFO nova.virt.driver [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Loading compute driver 'libvirt.LibvirtDriver'#033[00m Nov 27 04:47:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:47:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:47:06 localhost systemd[1]: session-60.scope: Deactivated successfully. Nov 27 04:47:06 localhost systemd[1]: session-60.scope: Consumed 1min 31.183s CPU time. Nov 27 04:47:06 localhost systemd-logind[761]: Session 60 logged out. Waiting for processes to exit. Nov 27 04:47:06 localhost systemd-logind[761]: Removed session 60. Nov 27 04:47:06 localhost systemd[1]: var-lib-containers-storage-overlay-16a331b99626a76219b5a3690d45b29d787d4868ae31d9393df1b3cc4b70afae-merged.mount: Deactivated successfully. Nov 27 04:47:06 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-67337598df6fbab15af30eac2db4dd5a049fd135d2f86383d0089ab30dd847ac-userdata-shm.mount: Deactivated successfully. Nov 27 04:47:06 localhost podman[284258]: 2025-11-27 09:47:06.230137306 +0000 UTC m=+0.071152837 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, config_id=edpm, release=1755695350, managed_by=edpm_ansible, version=9.6, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, architecture=x86_64, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, com.redhat.component=ubi9-minimal-container, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, maintainer=Red Hat, Inc., vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=) Nov 27 04:47:06 localhost podman[284258]: 2025-11-27 09:47:06.243898063 +0000 UTC m=+0.084913664 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, maintainer=Red Hat, Inc., version=9.6, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.expose-services=, architecture=x86_64, managed_by=edpm_ansible, io.openshift.tags=minimal rhel9, vcs-type=git, com.redhat.component=ubi9-minimal-container, release=1755695350, name=ubi9-minimal, config_id=edpm, build-date=2025-08-20T13:12:41, distribution-scope=public, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.261 284030 INFO nova.compute.provider_config [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access.#033[00m Nov 27 04:47:06 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.270 284030 DEBUG oslo_concurrency.lockutils [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.270 284030 DEBUG oslo_concurrency.lockutils [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.271 284030 DEBUG oslo_concurrency.lockutils [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.271 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Full set of CONF: _wait_for_exit_or_signal /usr/lib/python3.9/site-packages/oslo_service/service.py:362#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.271 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2589#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.271 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Configuration options gathered from: log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2590#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.271 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] command line args: [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2591#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2592#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ================================================================================ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2594#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] allow_resize_to_same_host = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] arq_binding_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] backdoor_port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] backdoor_socket = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] block_device_allocate_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.272 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] block_device_allocate_retries_interval = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cert = self.pem log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute_driver = libvirt.LibvirtDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute_monitors = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] config_dir = ['/etc/nova/nova.conf.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] config_drive_format = iso9660 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-compute.conf'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.273 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] config_source = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] console_host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] control_exchange = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cpu_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] debug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] default_access_ip_network_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] default_availability_zone = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.274 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] default_ephemeral_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] default_schedule_zone = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] disk_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] enable_new_services = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] enabled_apis = ['osapi_compute', 'metadata'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] enabled_ssl_apis = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.275 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] flat_injected = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] force_config_drive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] force_raw_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] graceful_shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] heal_instance_info_cache_interval = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] host = np0005537446.localdomain log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] initial_cpu_allocation_ratio = 4.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.276 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] initial_disk_allocation_ratio = 0.9 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] initial_ram_allocation_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] injected_network_template = /usr/lib/python3.9/site-packages/nova/virt/interfaces.template log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_build_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_delete_interval = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_name_template = instance-%08x log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_usage_audit = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.277 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_usage_audit_period = month log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instance_uuid_format = [instance: %(uuid)s] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] instances_path = /var/lib/nova/instances log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] internal_service_availability_zone = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] live_migration_retry_count = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_config_append = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.278 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_date_format = %Y-%m-%d %H:%M:%S log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_options = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_rotate_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_rotate_interval_type = days log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] log_rotation_type = size log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.279 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] logging_context_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(user_identity)s] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] logging_debug_format_suffix = %(funcName)s %(pathname)s:%(lineno)d log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] logging_default_format_string = %(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [-] %(instance)s%(message)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] logging_exception_prefix = %(asctime)s.%(msecs)03d %(process)d ERROR %(name)s %(instance)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] long_rpc_timeout = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.280 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_concurrent_builds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_concurrent_live_migrations = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_concurrent_snapshots = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_local_block_devices = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_logfile_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] max_logfile_size_mb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] maximum_instance_delete_attempts = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metadata_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.281 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metadata_listen_port = 8775 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metadata_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] migrate_max_retries = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] mkisofs_cmd = /usr/bin/mkisofs log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] my_block_storage_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] my_ip = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] network_allocate_retries = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.282 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] osapi_compute_listen = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] osapi_compute_listen_port = 8774 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] osapi_compute_unique_server_name_scope = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] osapi_compute_workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] password_length = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] periodic_enable = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost systemd[1]: tmp-crun.LRQhUG.mount: Deactivated successfully. Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] periodic_fuzzy_delay = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.283 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] pointer_model = usbtablet log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] preallocate_images = none log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] publish_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] pybasedir = /usr/lib/python3.9/site-packages log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ram_allocation_ratio = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rate_limit_burst = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rate_limit_except_level = CRITICAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rate_limit_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reboot_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.284 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reclaim_instance_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] record = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reimage_timeout_per_gb = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] report_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rescue_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reserved_host_cpus = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reserved_host_disk_mb = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reserved_host_memory_mb = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.285 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] reserved_huge_pages = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] resize_confirm_window = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] resize_fs_using_block_device = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] resume_guests_state_on_host_boot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rootwrap_config = /etc/nova/rootwrap.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rpc_response_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] run_external_periodic_tasks = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] running_deleted_instance_action = reap log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.286 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] running_deleted_instance_poll_interval = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] running_deleted_instance_timeout = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler_instance_sync_interval = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_down_time = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] servicegroup_driver = db log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] shelved_offload_time = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] shelved_poll_interval = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.287 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] shutdown_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] source_is_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ssl_only = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] state_path = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] sync_power_state_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] sync_power_state_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] syslog_log_facility = LOG_USER log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] tempdir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.288 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] timeout_nbd = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] update_resources_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_cow_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_eventlog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_journal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_json = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_rootwrap_daemon = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.289 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_stderr = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] use_syslog = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vcpu_pin_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plugging_is_fatal = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plugging_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] virt_mkfs = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] volume_usage_poll_interval = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] watch_log_file = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.290 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] web = /usr/share/spice-html5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2602#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_concurrency.disable_process_locking = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_concurrency.lock_path = /var/lib/nova/tmp log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_metrics.metrics_buffer_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_metrics.metrics_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_metrics.metrics_process_name = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.291 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.auth_strategy = keystone log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.compute_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.dhcp_domain = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.enable_instance_password = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.glance_link_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.instance_list_cells_batch_fixed_size = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.292 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.instance_list_cells_batch_strategy = distributed log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.instance_list_per_project_cells = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.list_records_by_skipping_down_cells = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.local_metadata_per_cell = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.max_limit = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.metadata_cache_expiration = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.neutron_default_tenant_id = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.293 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.use_forwarded_for = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.use_neutron_default_nets = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_dynamic_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_dynamic_failure_fatal = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_dynamic_read_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_dynamic_ssl_certfile = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_dynamic_targets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_jsonfile_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.294 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api.vendordata_providers = ['StaticJSON'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.backend = oslo_cache.dict log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.backend_argument = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.config_prefix = cache.oslo log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.dead_timeout = 60.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.debug_cache_backend = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.enable_retry_client = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.295 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.enable_socket_keepalive = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.expiration_time = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.hashclient_retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.hashclient_retry_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_dead_retry = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_password = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_pool_connection_get_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.296 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_pool_flush_on_reconnect = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_pool_maxsize = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_pool_unused_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost podman[284257]: 2025-11-27 09:47:06.296662939 +0000 UTC m=+0.142092568 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, config_id=edpm, org.label-schema.build-date=20251125) Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_sasl_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_servers = ['localhost:11211'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_socket_timeout = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.memcache_username = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.297 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.proxies = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.retry_attempts = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.retry_delay = 0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.socket_keepalive_count = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.socket_keepalive_idle = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.socket_keepalive_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.tls_allowed_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.tls_cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.298 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.tls_certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.tls_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cache.tls_keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.catalog_info = volumev3:cinderv3:internalURL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.299 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.cross_az_attach = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.endpoint_template = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.300 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.os_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cinder.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.consecutive_build_service_disable_threshold = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.cpu_dedicated_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.cpu_shared_set = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.image_type_exclude_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.live_migration_wait_for_vif_plug = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.301 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.max_concurrent_disk_ops = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.max_disk_devices_to_attach = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.packing_host_numa_cells_allocation_strategy = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.provider_config_location = /etc/nova/provider_config/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.resource_provider_association_refresh = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.shutdown_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.302 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] conductor.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] console.allowed_origins = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] console.ssl_ciphers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] console.ssl_minimum_version = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] consoleauth.token_ttl = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.303 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.304 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.service_type = accelerator log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.305 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] cyborg.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost podman[284257]: 2025-11-27 09:47:06.306792179 +0000 UTC m=+0.152221858 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_managed=true, org.label-schema.schema-version=1.0, config_id=edpm, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.306 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.307 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.backend = sqlalchemy log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.308 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.connection_debug = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.connection_parameters = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.connection_recycle_time = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.connection_trace = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.db_inc_retry_interval = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.db_max_retries = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.db_max_retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.309 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.db_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.max_overflow = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.max_pool_size = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.max_retries = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.mysql_enable_ndb = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.mysql_sql_mode = TRADITIONAL log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.mysql_wsrep_sync_wait = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.310 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.pool_timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.retry_interval = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.slave_connection = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] api_database.sqlite_synchronous = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] devices.enabled_mdev_types = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ephemeral_storage_encryption.cipher = aes-xts-plain64 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ephemeral_storage_encryption.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ephemeral_storage_encryption.key_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.311 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.api_servers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.312 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.default_trusted_certificate_ids = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.enable_certificate_validation = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.enable_rbd_download = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.313 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.num_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.rbd_ceph_conf = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.rbd_pool = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.rbd_user = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.service_type = image log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.314 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.verify_glance_signatures = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] glance.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.315 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] guestfs.debug = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.config_drive_cdrom = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.config_drive_inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.dynamic_memory_ratio = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.enable_instance_metrics_collection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.enable_remotefx = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.instances_path_share = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.iscsi_initiator_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.316 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.limit_cpu_features = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.mounted_disk_query_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.mounted_disk_query_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.power_state_check_timeframe = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.power_state_event_polling_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.qemu_img_cmd = qemu-img.exe log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.use_multipath_io = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.volume_attach_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.317 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.volume_attach_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.vswitch_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] hyperv.wait_soft_reboot_seconds = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] mks.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] mks.mksproxy_base_url = http://127.0.0.1:6090/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.manager_interval = 2400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.318 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.precache_concurrency = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.remove_unused_base_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.remove_unused_original_minimum_age_seconds = 86400 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.remove_unused_resized_minimum_age_seconds = 3600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] image_cache.subdirectory_name = _base log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.api_max_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.api_retry_interval = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.319 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.320 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.partition_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.peer_list = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.serial_console_state_timeout = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.321 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.service_type = baremetal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ironic.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.322 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] key_manager.backend = barbican log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] key_manager.fixed_key = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.auth_endpoint = http://localhost/identity/v3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.barbican_api_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.barbican_endpoint = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.barbican_endpoint_type = internal log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.barbican_region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.323 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.number_of_retries = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.retry_delay = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.send_service_user_token = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.324 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.verify_ssl = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican.verify_ssl_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.325 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] barbican_service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.approle_role_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.approle_secret_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.326 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.kv_mountpoint = secret log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.kv_version = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.327 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.root_token_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.use_ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vault.vault_url = http://127.0.0.1:8200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.328 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.329 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.service_type = identity log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.330 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.valid_interfaces = ['internal', 'public'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] keystone.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.connection_uri = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_mode = host-model log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_model_extra_flags = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_models = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_power_governor_high = performance log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_power_governor_low = powersave log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.331 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_power_management = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.cpu_power_management_strategy = cpu_state log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.device_detach_attempts = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.device_detach_timeout = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.disk_cachemodes = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.disk_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.enabled_perf_events = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.332 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.file_backed_memory = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.gid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.hw_disk_discard = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.hw_machine_type = ['x86_64=q35'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_rbd_ceph_conf = /etc/ceph/ceph.conf log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_rbd_glance_copy_poll_interval = 15 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_rbd_glance_copy_timeout = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_rbd_glance_store_name = default_backend log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.333 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_rbd_pool = vms log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_type = rbd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.images_volume_group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.inject_key = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.inject_partition = -2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.inject_password = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.iscsi_iface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.iser_use_multipath = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.334 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_bandwidth = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_completion_timeout = 800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_downtime = 500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_downtime_delay = 75 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_downtime_steps = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_inbound_addr = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_permit_auto_converge = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.335 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_permit_post_copy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_scheme = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_timeout_action = force_complete log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_tunnelled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 WARNING oslo_config.cfg [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] Deprecated: Option "live_migration_uri" from group "libvirt" is deprecated for removal ( Nov 27 04:47:06 localhost nova_compute[284026]: live_migration_uri is deprecated for removal in favor of two other options that Nov 27 04:47:06 localhost nova_compute[284026]: allow to change live migration scheme and target URI: ``live_migration_scheme`` Nov 27 04:47:06 localhost nova_compute[284026]: and ``live_migration_inbound_addr`` respectively. Nov 27 04:47:06 localhost nova_compute[284026]: ). Its value may be silently ignored in the future.#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_uri = qemu+ssh://nova@%s/system?keyfile=/var/lib/nova/.ssh/ssh-privatekey log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.live_migration_with_native_tls = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.336 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.max_queues = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.mem_stats_period_seconds = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.nfs_mount_options = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.nfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_aoe_discover_tries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_iser_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_memory_encrypted_guests = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.337 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_nvme_discover_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_pcie_ports = 24 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.num_volume_scan_tries = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.pmem_namespaces = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.quobyte_client_cfg = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.quobyte_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rbd_connect_timeout = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.338 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rbd_destroy_volume_retries = 12 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rbd_destroy_volume_retry_interval = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rbd_secret_uuid = e83f3b0c-4090-52df-95d4-ad9be8516692 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rbd_user = openstack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.realtime_scheduler_priority = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.remote_filesystem_transport = ssh log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rescue_image_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rescue_kernel_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.339 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rescue_ramdisk_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rng_dev_path = /dev/urandom log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.rx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.smbfs_mount_options = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.smbfs_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.snapshot_compression = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.snapshot_image_format = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.340 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.snapshots_directory = /var/lib/nova/instances/snapshots log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.sparse_logical_volumes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.swtpm_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.swtpm_group = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.swtpm_user = tss log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.sysinfo_serial = unique log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.tx_queue_size = 512 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.341 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.uid_maps = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.use_virtio_for_bridges = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.virt_type = kvm log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.volume_clear = zero log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.volume_clear_size = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.volume_use_multipath = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.342 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_cache_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.343 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.343 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_mount_group = qemu log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.343 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_mount_opts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.343 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_mount_perms = 0770 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_mount_point_base = /var/lib/nova/mnt log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.vzstorage_mount_user = stack log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] libvirt.wait_soft_reboot_seconds = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.344 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.default_floating_pool = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.extension_sync_interval = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.http_retries = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.345 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.metadata_proxy_shared_secret = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.ovs_bridge = br-int log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.physnets = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.346 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.service_metadata_proxy = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.service_type = network log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.347 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] neutron.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] notifications.bdms_in_notifications = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] notifications.default_level = INFO log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] notifications.notification_format = unversioned log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] notifications.notify_on_state_change = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] notifications.versioned_notifications_topics = ['versioned_notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.348 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] pci.alias = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] pci.device_spec = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] pci.report_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.349 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.350 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.project_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.351 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.project_name = service log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.region_name = regionOne log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.service_type = placement log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.352 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.system_scope = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.valid_interfaces = ['internal'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.353 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] placement.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.cores = 20 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.count_usage_from_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.driver = nova.quota.DbQuotaDriver log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.injected_file_content_bytes = 10240 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.injected_file_path_length = 255 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.injected_files = 5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.354 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.instances = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.key_pairs = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.metadata_items = 128 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.ram = 51200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.recheck_quota = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.server_group_members = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] quota.server_groups = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.355 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rdp.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] rdp.html5_proxy_base_url = http://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.discover_hosts_in_cells_interval = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.enable_isolated_aggregate_filtering = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.image_metadata_prefilter = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.limit_tenants_to_placement_aggregate = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.356 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.max_attempts = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.max_placement_results = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.placement_aggregate_required_for_tenants = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.query_placement_for_availability_zone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.query_placement_for_image_type_support = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.query_placement_for_routed_network_aggregates = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] scheduler.workers = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.357 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.aggregate_image_properties_isolation_namespace = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.aggregate_image_properties_isolation_separator = . log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.build_failure_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.cpu_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.disk_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.358 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.host_subset_size = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.image_properties_default_architecture = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.io_ops_weight_multiplier = -1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.isolated_hosts = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.isolated_images = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.max_instances_per_host = 50 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.max_io_ops_per_host = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.359 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.pci_in_placement = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.pci_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.ram_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.shuffle_best_same_weighed_hosts = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.soft_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.track_instance_changes = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.360 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metrics.required = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metrics.weight_multiplier = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metrics.weight_of_unavailable = -10000.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] metrics.weight_setting = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.base_url = ws://127.0.0.1:6083/ log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.361 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.port_range = 10000:20000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.serialproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] serial_console.serialproxy_port = 6083 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.362 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.send_service_user_token = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] service_user.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.agent_enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.363 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.html5proxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.html5proxy_port = 6082 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.image_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.jpeg_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.364 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.playback_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.server_listen = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.server_proxyclient_address = 127.0.0.1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.streaming_mode = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] spice.zlib_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] upgrade_levels.baseapi = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] upgrade_levels.cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] upgrade_levels.compute = auto log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.365 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] upgrade_levels.conductor = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] upgrade_levels.scheduler = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.auth_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.366 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vendordata_dynamic_auth.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.api_retry_count = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.cache_prefix = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.cluster_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.367 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.connection_pool_size = 10 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.console_delay_seconds = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.datastore_regex = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.host_ip = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.host_password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.host_port = 443 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.host_username = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.368 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.integration_bridge = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.maximum_objects = 100 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.pbm_default_policy = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.pbm_enabled = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.pbm_wsdl_location = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.serial_log_dir = /opt/vmware/vspc log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.serial_port_proxy_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.serial_port_service_uri = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.369 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.task_poll_interval = 0.5 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.use_linked_clone = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.vnc_keymap = en-us log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.vnc_port = 5900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vmware.vnc_port_total = 10000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.auth_schemes = ['none'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.370 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.enabled = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.novncproxy_base_url = http://nova-novncproxy-cell1-public-openstack.apps-crc.testing/vnc_lite.html log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.novncproxy_host = 0.0.0.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.novncproxy_port = 6080 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.server_listen = ::0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.server_proxyclient_address = 192.168.122.108 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.vencrypt_ca_certs = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.vencrypt_client_cert = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.371 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vnc.vencrypt_client_key = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_compute_service_check_for_ffu = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_deep_image_inspection = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_fallback_pcpu_query = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_group_policy_check_upcall = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_libvirt_livesnapshot = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.disable_rootwrap = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.372 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.enable_numa_live_migration = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.enable_qemu_monitor_announce_self = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.handle_virt_lifecycle_events = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.libvirt_disable_apic = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.never_download_image_if_on_rbd = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.qemu_monitor_announce_self_count = 3 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.qemu_monitor_announce_self_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.373 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.reserve_disk_resource_for_image_cache = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.skip_cpu_compare_at_startup = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.skip_cpu_compare_on_dest = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.skip_hypervisor_version_check_on_lm = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.skip_reserve_in_use_ironic_nodes = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.unified_limits_count_pcpu_as_vcpu = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.374 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.api_paste_config = api-paste.ini log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.client_socket_timeout = 900 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.default_pool_size = 1000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.keep_alive = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.max_header_line = 16384 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.secure_proxy_ssl_header = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.ssl_ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.ssl_cert_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.375 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.ssl_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.tcp_keepidle = 600 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] zvm.ca_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] zvm.cloud_connector_url = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] zvm.image_tmp_path = /var/lib/nova/images log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] zvm.reachable_timeout = 300 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.enforce_new_defaults = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.376 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.enforce_scope = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.policy_default_rule = default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.policy_dirs = ['policy.d'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.policy_file = policy.yaml log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.remote_content_type = application/x-www-form-urlencoded log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.remote_ssl_ca_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.remote_ssl_client_crt_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.377 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.remote_ssl_client_key_file = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_policy.remote_ssl_verify_server_crt = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_versionedobjects.fatal_exception_format_errors = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] remote_debug.host = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] remote_debug.port = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.amqp_auto_delete = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.378 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.amqp_durable_queues = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.conn_pool_min_size = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.conn_pool_ttl = 1200 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.direct_mandatory_flag = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.enable_cancel_on_failover = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.heartbeat_in_pthread = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.heartbeat_rate = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.379 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.kombu_compression = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.kombu_failover_strategy = round-robin log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_ha_queues = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_interval_max = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.380 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_quorum_queue = True log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_retry_backoff = 2 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_retry_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.381 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.rpc_conn_pool_size = 30 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl_ca_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl_cert_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl_enforce_fips_mode = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl_key_file = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.382 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_rabbit.ssl_version = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_notifications.driver = ['noop'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_notifications.retry = -1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_notifications.topics = ['notifications'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_messaging_notifications.transport_url = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.auth_section = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.auth_type = password log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.383 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.auth_url = http://keystone-internal.openstack.svc:5000 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.cafile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.certfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.collect_timing = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.connect_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.connect_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.default_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.default_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.384 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.endpoint_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.endpoint_override = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.insecure = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.keyfile = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.max_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.min_version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.385 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.password = **** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.project_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.project_domain_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.project_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.project_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.region_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.service_name = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.386 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.service_type = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.split_loggers = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.status_code_retries = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.status_code_retry_delay = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.system_scope = all log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.timeout = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.trust_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.user_domain_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.387 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.user_domain_name = Default log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.user_id = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.username = nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.valid_interfaces = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_limit.version = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_reports.file_event_handler = /var/lib/nova log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_reports.file_event_handler_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] oslo_reports.log_dir = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.388 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.capabilities = [12] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_linux_bridge_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.capabilities = [12, 1] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.389 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] vif_plug_ovs_privileged.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.flat_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.forward_bridge_interface = ['all'] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.iptables_bottom_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.390 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.iptables_drop_action = DROP log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.iptables_top_regex = log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.use_ipv6 = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_linux_bridge.vlan_interface = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.isolate_vif = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.network_device_mtu = 1500 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.391 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.ovs_vsctl_timeout = 120 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.ovsdb_interface = native log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_vif_ovs.per_port_bridge = False log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_brick.lock_path = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_brick.wait_mpath_device_attempts = 4 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] os_brick.wait_mpath_device_interval = 1 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.capabilities = [21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.392 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.logger_name = os_brick.privileged log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] privsep_osbrick.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.group = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.helper_command = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.393 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.logger_name = oslo_privsep.daemon log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.394 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.thread_pool_size = 8 log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.394 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] nova_sys_admin.user = None log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2609#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.394 284030 DEBUG oslo_service.service [None req-77e32036-9e44-4bb0-9160-7580c0407989 - - - - - -] ******************************************************************************** log_opt_values /usr/lib/python3.9/site-packages/oslo_config/cfg.py:2613#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.395 284030 INFO nova.service [-] Starting compute node (version 27.5.2-0.20250829104910.6f8decf.el9)#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.408 284030 INFO nova.virt.node [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.408 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Starting native event thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:492#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.409 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Starting green dispatch thread _init_events /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:498#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.409 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Starting connection event dispatch thread initialize /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:620#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.409 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Connecting to libvirt: qemu:///system _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:503#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.419 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Registering for lifecycle events _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:509#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.421 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Registering for connection events: _get_new_connection /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:530#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.421 284030 INFO nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Connection event '1' reason 'None'#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.427 284030 INFO nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Libvirt host capabilities Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 8a8a8082-6126-4917-bb42-c4150a0ac6a5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: x86_64 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v4 Nov 27 04:47:06 localhost nova_compute[284026]: AMD Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tcp Nov 27 04:47:06 localhost nova_compute[284026]: rdma Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 16116612 Nov 27 04:47:06 localhost nova_compute[284026]: 4029153 Nov 27 04:47:06 localhost nova_compute[284026]: 0 Nov 27 04:47:06 localhost nova_compute[284026]: 0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: selinux Nov 27 04:47:06 localhost nova_compute[284026]: 0 Nov 27 04:47:06 localhost nova_compute[284026]: system_u:system_r:svirt_t:s0 Nov 27 04:47:06 localhost nova_compute[284026]: system_u:system_r:svirt_tcg_t:s0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: dac Nov 27 04:47:06 localhost nova_compute[284026]: 0 Nov 27 04:47:06 localhost nova_compute[284026]: +107:+107 Nov 27 04:47:06 localhost nova_compute[284026]: +107:+107 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: hvm Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 32 Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-i440fx-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.8.0 Nov 27 04:47:06 localhost nova_compute[284026]: q35 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.4.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.5.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.3.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.4.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.2.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.2.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.0.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.0.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.1.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: hvm Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 64 Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-i440fx-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.8.0 Nov 27 04:47:06 localhost nova_compute[284026]: q35 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.4.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.5.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.3.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.4.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.2.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.2.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.0.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.0.0 Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel8.1.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: #033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.434 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Getting domain capabilities for i686 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.437 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=pc: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-i440fx-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: i686 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: rom Nov 27 04:47:06 localhost nova_compute[284026]: pflash Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: yes Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: AMD Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 486 Nov 27 04:47:06 localhost nova_compute[284026]: 486-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Conroe Nov 27 04:47:06 localhost nova_compute[284026]: Conroe-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-IBPB Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v4 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v1 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v2 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v6 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v7 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Penryn Nov 27 04:47:06 localhost nova_compute[284026]: Penryn-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Westmere Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v2 Nov 27 04:47:06 localhost nova_compute[284026]: athlon Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: athlon-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: kvm32 Nov 27 04:47:06 localhost nova_compute[284026]: kvm32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: n270 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: n270-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pentium Nov 27 04:47:06 localhost nova_compute[284026]: pentium-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: phenom Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: phenom-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu32 Nov 27 04:47:06 localhost nova_compute[284026]: qemu32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: anonymous Nov 27 04:47:06 localhost nova_compute[284026]: memfd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: disk Nov 27 04:47:06 localhost nova_compute[284026]: cdrom Nov 27 04:47:06 localhost nova_compute[284026]: floppy Nov 27 04:47:06 localhost nova_compute[284026]: lun Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: ide Nov 27 04:47:06 localhost nova_compute[284026]: fdc Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: sata Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: vnc Nov 27 04:47:06 localhost nova_compute[284026]: egl-headless Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: subsystem Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: mandatory Nov 27 04:47:06 localhost nova_compute[284026]: requisite Nov 27 04:47:06 localhost nova_compute[284026]: optional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: pci Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: random Nov 27 04:47:06 localhost nova_compute[284026]: egd Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: path Nov 27 04:47:06 localhost nova_compute[284026]: handle Nov 27 04:47:06 localhost nova_compute[284026]: virtiofs Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tpm-tis Nov 27 04:47:06 localhost nova_compute[284026]: tpm-crb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: emulator Nov 27 04:47:06 localhost nova_compute[284026]: external Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 2.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: passt Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: isa Nov 27 04:47:06 localhost nova_compute[284026]: hyperv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: null Nov 27 04:47:06 localhost nova_compute[284026]: vc Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: dev Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: pipe Nov 27 04:47:06 localhost nova_compute[284026]: stdio Nov 27 04:47:06 localhost nova_compute[284026]: udp Nov 27 04:47:06 localhost nova_compute[284026]: tcp Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: qemu-vdagent Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: relaxed Nov 27 04:47:06 localhost nova_compute[284026]: vapic Nov 27 04:47:06 localhost nova_compute[284026]: spinlocks Nov 27 04:47:06 localhost nova_compute[284026]: vpindex Nov 27 04:47:06 localhost nova_compute[284026]: runtime Nov 27 04:47:06 localhost nova_compute[284026]: synic Nov 27 04:47:06 localhost nova_compute[284026]: stimer Nov 27 04:47:06 localhost nova_compute[284026]: reset Nov 27 04:47:06 localhost nova_compute[284026]: vendor_id Nov 27 04:47:06 localhost nova_compute[284026]: frequencies Nov 27 04:47:06 localhost nova_compute[284026]: reenlightenment Nov 27 04:47:06 localhost nova_compute[284026]: tlbflush Nov 27 04:47:06 localhost nova_compute[284026]: ipi Nov 27 04:47:06 localhost nova_compute[284026]: avic Nov 27 04:47:06 localhost nova_compute[284026]: emsr_bitmap Nov 27 04:47:06 localhost nova_compute[284026]: xmm_input Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 4095 Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Linux KVM Hv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tdx Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.441 284030 DEBUG nova.virt.libvirt.volume.mount [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Initialising _HostMountState generation 0 host_up /usr/lib/python3.9/site-packages/nova/virt/libvirt/volume/mount.py:130#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.443 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Libvirt host hypervisor capabilities for arch=i686 and machine_type=q35: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.8.0 Nov 27 04:47:06 localhost nova_compute[284026]: i686 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: rom Nov 27 04:47:06 localhost nova_compute[284026]: pflash Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: yes Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: AMD Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 486 Nov 27 04:47:06 localhost nova_compute[284026]: 486-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Conroe Nov 27 04:47:06 localhost nova_compute[284026]: Conroe-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-IBPB Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v4 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v1 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v2 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v6 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v7 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Penryn Nov 27 04:47:06 localhost nova_compute[284026]: Penryn-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Westmere Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v2 Nov 27 04:47:06 localhost nova_compute[284026]: athlon Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: athlon-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: kvm32 Nov 27 04:47:06 localhost nova_compute[284026]: kvm32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: n270 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: n270-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pentium Nov 27 04:47:06 localhost nova_compute[284026]: pentium-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: phenom Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: phenom-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu32 Nov 27 04:47:06 localhost nova_compute[284026]: qemu32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: anonymous Nov 27 04:47:06 localhost nova_compute[284026]: memfd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: disk Nov 27 04:47:06 localhost nova_compute[284026]: cdrom Nov 27 04:47:06 localhost nova_compute[284026]: floppy Nov 27 04:47:06 localhost nova_compute[284026]: lun Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: fdc Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: sata Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: vnc Nov 27 04:47:06 localhost nova_compute[284026]: egl-headless Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: subsystem Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: mandatory Nov 27 04:47:06 localhost nova_compute[284026]: requisite Nov 27 04:47:06 localhost nova_compute[284026]: optional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: pci Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: random Nov 27 04:47:06 localhost nova_compute[284026]: egd Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: path Nov 27 04:47:06 localhost nova_compute[284026]: handle Nov 27 04:47:06 localhost nova_compute[284026]: virtiofs Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tpm-tis Nov 27 04:47:06 localhost nova_compute[284026]: tpm-crb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: emulator Nov 27 04:47:06 localhost nova_compute[284026]: external Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 2.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: passt Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: isa Nov 27 04:47:06 localhost nova_compute[284026]: hyperv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: null Nov 27 04:47:06 localhost nova_compute[284026]: vc Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: dev Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: pipe Nov 27 04:47:06 localhost nova_compute[284026]: stdio Nov 27 04:47:06 localhost nova_compute[284026]: udp Nov 27 04:47:06 localhost nova_compute[284026]: tcp Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: qemu-vdagent Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: relaxed Nov 27 04:47:06 localhost nova_compute[284026]: vapic Nov 27 04:47:06 localhost nova_compute[284026]: spinlocks Nov 27 04:47:06 localhost nova_compute[284026]: vpindex Nov 27 04:47:06 localhost nova_compute[284026]: runtime Nov 27 04:47:06 localhost nova_compute[284026]: synic Nov 27 04:47:06 localhost nova_compute[284026]: stimer Nov 27 04:47:06 localhost nova_compute[284026]: reset Nov 27 04:47:06 localhost nova_compute[284026]: vendor_id Nov 27 04:47:06 localhost nova_compute[284026]: frequencies Nov 27 04:47:06 localhost nova_compute[284026]: reenlightenment Nov 27 04:47:06 localhost nova_compute[284026]: tlbflush Nov 27 04:47:06 localhost nova_compute[284026]: ipi Nov 27 04:47:06 localhost nova_compute[284026]: avic Nov 27 04:47:06 localhost nova_compute[284026]: emsr_bitmap Nov 27 04:47:06 localhost nova_compute[284026]: xmm_input Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 4095 Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Linux KVM Hv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tdx Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.473 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Getting domain capabilities for x86_64 via machine types: {'pc', 'q35'} _get_machine_types /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:952#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.484 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=pc: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-i440fx-rhel7.6.0 Nov 27 04:47:06 localhost nova_compute[284026]: x86_64 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/OVMF/OVMF_CODE.secboot.fd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: rom Nov 27 04:47:06 localhost nova_compute[284026]: pflash Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: yes Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: AMD Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 486 Nov 27 04:47:06 localhost nova_compute[284026]: 486-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Conroe Nov 27 04:47:06 localhost nova_compute[284026]: Conroe-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-IBPB Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v4 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v1 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v2 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v6 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v7 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Penryn Nov 27 04:47:06 localhost nova_compute[284026]: Penryn-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Westmere Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v2 Nov 27 04:47:06 localhost nova_compute[284026]: athlon Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: athlon-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: kvm32 Nov 27 04:47:06 localhost nova_compute[284026]: kvm32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: n270 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: n270-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pentium Nov 27 04:47:06 localhost nova_compute[284026]: pentium-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: phenom Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: phenom-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu32 Nov 27 04:47:06 localhost nova_compute[284026]: qemu32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: anonymous Nov 27 04:47:06 localhost nova_compute[284026]: memfd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: disk Nov 27 04:47:06 localhost nova_compute[284026]: cdrom Nov 27 04:47:06 localhost nova_compute[284026]: floppy Nov 27 04:47:06 localhost nova_compute[284026]: lun Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: ide Nov 27 04:47:06 localhost nova_compute[284026]: fdc Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: sata Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: vnc Nov 27 04:47:06 localhost nova_compute[284026]: egl-headless Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: subsystem Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: mandatory Nov 27 04:47:06 localhost nova_compute[284026]: requisite Nov 27 04:47:06 localhost nova_compute[284026]: optional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: pci Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: random Nov 27 04:47:06 localhost nova_compute[284026]: egd Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: path Nov 27 04:47:06 localhost nova_compute[284026]: handle Nov 27 04:47:06 localhost nova_compute[284026]: virtiofs Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tpm-tis Nov 27 04:47:06 localhost nova_compute[284026]: tpm-crb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: emulator Nov 27 04:47:06 localhost nova_compute[284026]: external Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 2.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: passt Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: isa Nov 27 04:47:06 localhost nova_compute[284026]: hyperv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: null Nov 27 04:47:06 localhost nova_compute[284026]: vc Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: dev Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: pipe Nov 27 04:47:06 localhost nova_compute[284026]: stdio Nov 27 04:47:06 localhost nova_compute[284026]: udp Nov 27 04:47:06 localhost nova_compute[284026]: tcp Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: qemu-vdagent Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: relaxed Nov 27 04:47:06 localhost nova_compute[284026]: vapic Nov 27 04:47:06 localhost nova_compute[284026]: spinlocks Nov 27 04:47:06 localhost nova_compute[284026]: vpindex Nov 27 04:47:06 localhost nova_compute[284026]: runtime Nov 27 04:47:06 localhost nova_compute[284026]: synic Nov 27 04:47:06 localhost nova_compute[284026]: stimer Nov 27 04:47:06 localhost nova_compute[284026]: reset Nov 27 04:47:06 localhost nova_compute[284026]: vendor_id Nov 27 04:47:06 localhost nova_compute[284026]: frequencies Nov 27 04:47:06 localhost nova_compute[284026]: reenlightenment Nov 27 04:47:06 localhost nova_compute[284026]: tlbflush Nov 27 04:47:06 localhost nova_compute[284026]: ipi Nov 27 04:47:06 localhost nova_compute[284026]: avic Nov 27 04:47:06 localhost nova_compute[284026]: emsr_bitmap Nov 27 04:47:06 localhost nova_compute[284026]: xmm_input Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 4095 Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Linux KVM Hv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tdx Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.529 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Libvirt host hypervisor capabilities for arch=x86_64 and machine_type=q35: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/libexec/qemu-kvm Nov 27 04:47:06 localhost nova_compute[284026]: kvm Nov 27 04:47:06 localhost nova_compute[284026]: pc-q35-rhel9.8.0 Nov 27 04:47:06 localhost nova_compute[284026]: x86_64 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: efi Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/edk2/ovmf/OVMF_CODE.secboot.fd Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/edk2/ovmf/OVMF_CODE.fd Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/edk2/ovmf/OVMF.amdsev.fd Nov 27 04:47:06 localhost nova_compute[284026]: /usr/share/edk2/ovmf/OVMF.inteltdx.secboot.fd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: rom Nov 27 04:47:06 localhost nova_compute[284026]: pflash Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: yes Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: yes Nov 27 04:47:06 localhost nova_compute[284026]: no Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: AMD Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 486 Nov 27 04:47:06 localhost nova_compute[284026]: 486-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Broadwell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cascadelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Conroe Nov 27 04:47:06 localhost nova_compute[284026]: Conroe-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Cooperlake-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Denverton-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Dhyana-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Genoa-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-IBPB Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Milan-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-Rome-v4 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v1 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v2 Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: EPYC-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: GraniteRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Haswell-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-noTSX Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v6 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Icelake-Server-v7 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: IvyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: KnightsMill-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nehalem-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G1-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G4-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Opteron_G5-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Penryn Nov 27 04:47:06 localhost nova_compute[284026]: Penryn-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: SandyBridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SapphireRapids-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: SierraForest-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Client-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-noTSX-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Skylake-Server-v5 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v2 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v3 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Snowridge-v4 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Westmere Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-IBRS Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Westmere-v2 Nov 27 04:47:06 localhost nova_compute[284026]: athlon Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: athlon-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: core2duo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: coreduo-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: kvm32 Nov 27 04:47:06 localhost nova_compute[284026]: kvm32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64 Nov 27 04:47:06 localhost nova_compute[284026]: kvm64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: n270 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: n270-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pentium Nov 27 04:47:06 localhost nova_compute[284026]: pentium-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2 Nov 27 04:47:06 localhost nova_compute[284026]: pentium2-v1 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3 Nov 27 04:47:06 localhost nova_compute[284026]: pentium3-v1 Nov 27 04:47:06 localhost nova_compute[284026]: phenom Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: phenom-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu32 Nov 27 04:47:06 localhost nova_compute[284026]: qemu32-v1 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64 Nov 27 04:47:06 localhost nova_compute[284026]: qemu64-v1 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: anonymous Nov 27 04:47:06 localhost nova_compute[284026]: memfd Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: disk Nov 27 04:47:06 localhost nova_compute[284026]: cdrom Nov 27 04:47:06 localhost nova_compute[284026]: floppy Nov 27 04:47:06 localhost nova_compute[284026]: lun Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: fdc Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: sata Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: vnc Nov 27 04:47:06 localhost nova_compute[284026]: egl-headless Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: subsystem Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: mandatory Nov 27 04:47:06 localhost nova_compute[284026]: requisite Nov 27 04:47:06 localhost nova_compute[284026]: optional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: pci Nov 27 04:47:06 localhost nova_compute[284026]: scsi Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: virtio Nov 27 04:47:06 localhost nova_compute[284026]: virtio-transitional Nov 27 04:47:06 localhost nova_compute[284026]: virtio-non-transitional Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: random Nov 27 04:47:06 localhost nova_compute[284026]: egd Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: path Nov 27 04:47:06 localhost nova_compute[284026]: handle Nov 27 04:47:06 localhost nova_compute[284026]: virtiofs Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tpm-tis Nov 27 04:47:06 localhost nova_compute[284026]: tpm-crb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: emulator Nov 27 04:47:06 localhost nova_compute[284026]: external Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 2.0 Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: usb Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: qemu Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: builtin Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: default Nov 27 04:47:06 localhost nova_compute[284026]: passt Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: isa Nov 27 04:47:06 localhost nova_compute[284026]: hyperv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: null Nov 27 04:47:06 localhost nova_compute[284026]: vc Nov 27 04:47:06 localhost nova_compute[284026]: pty Nov 27 04:47:06 localhost nova_compute[284026]: dev Nov 27 04:47:06 localhost nova_compute[284026]: file Nov 27 04:47:06 localhost nova_compute[284026]: pipe Nov 27 04:47:06 localhost nova_compute[284026]: stdio Nov 27 04:47:06 localhost nova_compute[284026]: udp Nov 27 04:47:06 localhost nova_compute[284026]: tcp Nov 27 04:47:06 localhost nova_compute[284026]: unix Nov 27 04:47:06 localhost nova_compute[284026]: qemu-vdagent Nov 27 04:47:06 localhost nova_compute[284026]: dbus Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: relaxed Nov 27 04:47:06 localhost nova_compute[284026]: vapic Nov 27 04:47:06 localhost nova_compute[284026]: spinlocks Nov 27 04:47:06 localhost nova_compute[284026]: vpindex Nov 27 04:47:06 localhost nova_compute[284026]: runtime Nov 27 04:47:06 localhost nova_compute[284026]: synic Nov 27 04:47:06 localhost nova_compute[284026]: stimer Nov 27 04:47:06 localhost nova_compute[284026]: reset Nov 27 04:47:06 localhost nova_compute[284026]: vendor_id Nov 27 04:47:06 localhost nova_compute[284026]: frequencies Nov 27 04:47:06 localhost nova_compute[284026]: reenlightenment Nov 27 04:47:06 localhost nova_compute[284026]: tlbflush Nov 27 04:47:06 localhost nova_compute[284026]: ipi Nov 27 04:47:06 localhost nova_compute[284026]: avic Nov 27 04:47:06 localhost nova_compute[284026]: emsr_bitmap Nov 27 04:47:06 localhost nova_compute[284026]: xmm_input Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: 4095 Nov 27 04:47:06 localhost nova_compute[284026]: on Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: off Nov 27 04:47:06 localhost nova_compute[284026]: Linux KVM Hv Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: tdx Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: Nov 27 04:47:06 localhost nova_compute[284026]: _get_domain_capabilities /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1037#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.583 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.584 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.584 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Checking secure boot support for host arch (x86_64) supports_secure_boot /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1782#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.584 284030 INFO nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Secure Boot support detected#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.586 284030 INFO nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.586 284030 INFO nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] The live_migration_permit_post_copy is set to True and post copy live migration is available so auto-converge will not be in use.#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.599 284030 DEBUG nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Enabling emulated TPM support _check_vtpm_support /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:1097#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.658 284030 INFO nova.virt.node [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Determined node identity 5764ad4c-cf89-4d5f-a185-92be71f7b67f from /var/lib/nova/compute_id#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.689 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Verified node 5764ad4c-cf89-4d5f-a185-92be71f7b67f matches my host np0005537446.localdomain _check_for_host_rename /usr/lib/python3.9/site-packages/nova/compute/manager.py:1568#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.740 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.744 284030 DEBUG nova.virt.libvirt.vif [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=,hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=,node='np0005537446.localdomain',numa_topology=None,old_flavor=,os_type=None,pci_devices=,pci_requests=,power_state=1,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata=,tags=,task_state=None,terminated_at=None,trusted_certs=,updated_at=2025-11-27T08:33:11Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.745 284030 DEBUG nova.network.os_vif_util [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.746 284030 DEBUG nova.network.os_vif_util [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Converted object VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.747 284030 DEBUG os_vif [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Plugging vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.829 284030 DEBUG ovsdbapp.backend.ovs_idl [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Created schema index Interface.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.829 284030 DEBUG ovsdbapp.backend.ovs_idl [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Created schema index Port.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.829 284030 DEBUG ovsdbapp.backend.ovs_idl [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Created schema index Bridge.name autocreate_indices /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.830 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] tcp:127.0.0.1:6640: entering CONNECTING _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.830 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [POLLOUT] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.831 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.831 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.833 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.837 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.863 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.864 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.864 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:47:06 localhost nova_compute[284026]: 2025-11-27 09:47:06.865 284030 INFO oslo.privsep.daemon [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Running privsep helper: ['sudo', 'nova-rootwrap', '/etc/nova/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-compute.conf', '--config-dir', '/etc/nova/nova.conf.d', '--privsep_context', 'vif_plug_ovs.privsep.vif_plug', '--privsep_sock_path', '/tmp/tmpuw4v4gof/privsep.sock']#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.458 284030 INFO oslo.privsep.daemon [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Spawned new privsep daemon via rootwrap#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.357 284319 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.362 284319 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.365 284319 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_DAC_OVERRIDE|CAP_NET_ADMIN/CAP_DAC_OVERRIDE|CAP_NET_ADMIN/none#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.365 284319 INFO oslo.privsep.daemon [-] privsep daemon running as pid 284319#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.746 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.747 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tapa2718872-3b, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.747 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tapa2718872-3b, col_values=(('external_ids', {'iface-id': 'a2718872-3b82-44ca-8d67-a0e516b2708b', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:86:b9:b3', 'vm-uuid': 'a02f7f2f-d2cf-4612-b18a-8be435257201'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.749 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.749 284030 INFO os_vif [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Successfully plugged vif VIFOpenVSwitch(active=True,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b')#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.750 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.754 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Current state is 1, state in DB is 1. _init_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:1304#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.755 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.890 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.891 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.891 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.892 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:47:07 localhost nova_compute[284026]: 2025-11-27 09:47:07.892 284030 DEBUG oslo_concurrency.processutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.306 284030 DEBUG oslo_concurrency.processutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.413s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.366 284030 DEBUG nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.366 284030 DEBUG nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.596 284030 WARNING nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.597 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11831MB free_disk=41.83721923828125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.598 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.598 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.743 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.744 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.744 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.758 284030 DEBUG nova.scheduler.client.report [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:47:08 localhost podman[242678]: time="2025-11-27T09:47:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:47:08 localhost podman[242678]: @ - - [27/Nov/2025:09:47:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.802 284030 DEBUG nova.scheduler.client.report [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.802 284030 DEBUG nova.compute.provider_tree [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:47:08 localhost podman[242678]: @ - - [27/Nov/2025:09:47:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18197 "" "Go-http-client/1.1" Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.819 284030 DEBUG nova.scheduler.client.report [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.837 284030 DEBUG nova.scheduler.client.report [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:47:08 localhost nova_compute[284026]: 2025-11-27 09:47:08.870 284030 DEBUG oslo_concurrency.processutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.365 284030 DEBUG oslo_concurrency.processutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.496s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.372 284030 DEBUG nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] /sys/module/kvm_amd/parameters/sev contains [N Nov 27 04:47:09 localhost nova_compute[284026]: ] _kernel_supports_amd_sev /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1803#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.372 284030 INFO nova.virt.libvirt.host [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] kernel doesn't support AMD SEV#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.374 284030 DEBUG nova.compute.provider_tree [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.374 284030 DEBUG nova.virt.libvirt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] CPU mode 'host-model' models '' was chosen, with extra flags: '' _get_guest_cpu_model_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:5396#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.399 284030 DEBUG nova.scheduler.client.report [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.424 284030 DEBUG nova.compute.resource_tracker [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.424 284030 DEBUG oslo_concurrency.lockutils [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.826s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.424 284030 DEBUG nova.service [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Creating RPC server for service compute start /usr/lib/python3.9/site-packages/nova/service.py:182#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.456 284030 DEBUG nova.service [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Join ServiceGroup membership for this service compute start /usr/lib/python3.9/site-packages/nova/service.py:199#033[00m Nov 27 04:47:09 localhost nova_compute[284026]: 2025-11-27 09:47:09.457 284030 DEBUG nova.servicegroup.drivers.db [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] DB_Driver: join new ServiceGroup member np0005537446.localdomain to the compute group, service = join /usr/lib/python3.9/site-packages/nova/servicegroup/drivers/db.py:44#033[00m Nov 27 04:47:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:47:09 localhost podman[284367]: 2025-11-27 09:47:09.98995922 +0000 UTC m=+0.083605539 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:47:10 localhost podman[284367]: 2025-11-27 09:47:10.001892218 +0000 UTC m=+0.095538557 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:47:10 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:47:11 localhost nova_compute[284026]: 2025-11-27 09:47:11.753 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:11 localhost nova_compute[284026]: 2025-11-27 09:47:11.835 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40975 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A017B90000000001030307) Nov 27 04:47:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40976 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A01BDA0000000001030307) Nov 27 04:47:16 localhost nova_compute[284026]: 2025-11-27 09:47:16.756 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:16 localhost nova_compute[284026]: 2025-11-27 09:47:16.836 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52477 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A01DDA0000000001030307) Nov 27 04:47:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:47:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:47:17 localhost podman[284391]: 2025-11-27 09:47:17.981738422 +0000 UTC m=+0.078991826 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:47:18 localhost podman[284392]: 2025-11-27 09:47:18.041821803 +0000 UTC m=+0.136898669 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:47:18 localhost podman[284392]: 2025-11-27 09:47:18.050211317 +0000 UTC m=+0.145288123 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:47:18 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:47:18 localhost podman[284391]: 2025-11-27 09:47:18.079929989 +0000 UTC m=+0.177183353 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_managed=true, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:47:18 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:47:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40977 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A023DB0000000001030307) Nov 27 04:47:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=39932 DF PROTO=TCP SPT=35350 DPT=9102 SEQ=1021629493 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A027DB0000000001030307) Nov 27 04:47:21 localhost nova_compute[284026]: 2025-11-27 09:47:21.810 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:21 localhost nova_compute[284026]: 2025-11-27 09:47:21.838 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40978 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A0339A0000000001030307) Nov 27 04:47:25 localhost openstack_network_exporter[244641]: ERROR 09:47:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:47:25 localhost openstack_network_exporter[244641]: ERROR 09:47:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:47:25 localhost openstack_network_exporter[244641]: ERROR 09:47:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:47:25 localhost openstack_network_exporter[244641]: ERROR 09:47:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:47:25 localhost openstack_network_exporter[244641]: Nov 27 04:47:25 localhost openstack_network_exporter[244641]: ERROR 09:47:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:47:25 localhost openstack_network_exporter[244641]: Nov 27 04:47:26 localhost nova_compute[284026]: 2025-11-27 09:47:26.349 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:26.349 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=6, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=5) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 04:47:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:26.351 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 1 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 04:47:26 localhost nova_compute[284026]: 2025-11-27 09:47:26.841 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:27 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:27.353 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '6'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.459 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.480 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.481 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.482 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.482 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_running_deleted_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:47:27 localhost nova_compute[284026]: 2025-11-27 09:47:27.523 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.042s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:30 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40979 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A053DA0000000001030307) Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.845 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4999-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.848 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.849 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5004 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.849 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.876 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:31 localhost nova_compute[284026]: 2025-11-27 09:47:31.877 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:47:31 localhost podman[284438]: 2025-11-27 09:47:31.983788663 +0000 UTC m=+0.075520843 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, config_id=multipathd, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 04:47:32 localhost podman[284438]: 2025-11-27 09:47:32.026005239 +0000 UTC m=+0.117737399 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_id=multipathd, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:47:32 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:47:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:47:34 localhost systemd[1]: tmp-crun.3XlUDA.mount: Deactivated successfully. Nov 27 04:47:34 localhost podman[284458]: 2025-11-27 09:47:34.009627415 +0000 UTC m=+0.099024421 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:47:34 localhost podman[284458]: 2025-11-27 09:47:34.039233763 +0000 UTC m=+0.128630719 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2) Nov 27 04:47:34 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:47:36 localhost nova_compute[284026]: 2025-11-27 09:47:36.878 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4996-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:47:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:47:36 localhost systemd[1]: tmp-crun.GFMnoO.mount: Deactivated successfully. Nov 27 04:47:37 localhost podman[284477]: 2025-11-27 09:47:37.000760983 +0000 UTC m=+0.089692442 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-08-20T13:12:41, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, maintainer=Red Hat, Inc., architecture=x86_64, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, name=ubi9-minimal, config_id=edpm) Nov 27 04:47:37 localhost podman[284476]: 2025-11-27 09:47:36.971313947 +0000 UTC m=+0.065891476 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0) Nov 27 04:47:37 localhost podman[284477]: 2025-11-27 09:47:37.042085383 +0000 UTC m=+0.131016792 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, version=9.6, release=1755695350, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, vcs-type=git, build-date=2025-08-20T13:12:41, config_id=edpm, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, distribution-scope=public, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:47:37 localhost podman[284476]: 2025-11-27 09:47:37.051733931 +0000 UTC m=+0.146311460 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, managed_by=edpm_ansible, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:47:37 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:47:37 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:47:38 localhost podman[242678]: time="2025-11-27T09:47:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:47:38 localhost podman[242678]: @ - - [27/Nov/2025:09:47:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:47:38 localhost podman[242678]: @ - - [27/Nov/2025:09:47:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18203 "" "Go-http-client/1.1" Nov 27 04:47:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:47:40 localhost podman[284516]: 2025-11-27 09:47:40.997233424 +0000 UTC m=+0.083968889 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:47:41 localhost podman[284516]: 2025-11-27 09:47:41.008926705 +0000 UTC m=+0.095662210 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:47:41 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.885 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4999-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.887 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.888 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.888 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.906 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:41 localhost nova_compute[284026]: 2025-11-27 09:47:41.906 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:43.555 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:43.555 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:43.556 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:43 localhost nova_compute[284026]: 2025-11-27 09:47:43.675 284030 DEBUG nova.compute.manager [None req-8b2abe0e-c8f6-448c-9c90-1a3360d9c0a2 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:47:43 localhost nova_compute[284026]: 2025-11-27 09:47:43.679 284030 INFO nova.compute.manager [None req-8b2abe0e-c8f6-448c-9c90-1a3360d9c0a2 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Retrieving diagnostics#033[00m Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.132 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.133 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.136 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 434 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '46dd3701-974d-45a4-bd45-4475dd7f71f4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 434, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.133438', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f5eb7f0-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'ea449de909a710832a089cc220c44aaf0b0f7237715069a21340829a187e6457'}]}, 'timestamp': '2025-11-27 09:47:44.136886', '_unique_id': '153bc91b3cbd417cb069bb2de87cdf5f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.138 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.139 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.139 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.139 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 9789 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2ad4b3b8-1545-4c2b-9729-1e87e1aacb13', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9789, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.139414', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f5f2dca-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'ce02bbb8531385cb59103c1dddff464c8d4c785c9ff8bf512e898fd510d23d2e'}]}, 'timestamp': '2025-11-27 09:47:44.139821', '_unique_id': '3eba5884a0a24b208c25967678f68968'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.140 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.141 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.141 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '264f9946-9270-40ff-a4d7-d066e58b93c5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.141604', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f5f82ac-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': '25f8edb52c06b34dbf856ccac4a35f3c6442af8cbe7e288502610fc09e020c9d'}]}, 'timestamp': '2025-11-27 09:47:44.141982', '_unique_id': '220394c70f2a4365a31ff0fe15c92f96'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.143 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.156 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.157 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f44adb7d-b09f-40c7-afe5-3c710a29925c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.143807', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f61c7b0-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': '6b30f3bf77f93c83b545901bf731f4751723cc022a5f59163954829a5c721f45'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.143807', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f61e290-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': 'ee677e2f91739c5559a00d1706fdd3543599546bf565072c438768053dfc32e9'}]}, 'timestamp': '2025-11-27 09:47:44.157695', '_unique_id': '2b7798b6e35c411aa07daa29cbd8933b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.159 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.161 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.164 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e289ba67-eabe-4381-8c7a-f1381f08137d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.163854', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f63059e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': '00bb80386b102c08c52997b4394e780de6bf49792c94e62e793f2fcb5283c90f'}]}, 'timestamp': '2025-11-27 09:47:44.165246', '_unique_id': 'b0a0720e28294ce887b5f07b41eee91b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.168 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.201 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 974916304 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.201 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 177343418 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '325f6485-8499-4742-a959-724cf7355961', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 974916304, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.168786', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f68a710-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': 'edc48ada39c7611ff41700b7d8af6daca06874bd159454004931d26f0f3c9ca8'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 177343418, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.168786', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f68b8d6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '7edf5caa2236b0528907866ab5e3d52af9dd4981124e9ac09273a8138446eb52'}]}, 'timestamp': '2025-11-27 09:47:44.202365', '_unique_id': '129475254f9840b7bddf5dab4de602b6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.204 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.205 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.205 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'aa109ec7-5045-4e0c-98ce-17e16ce6c71e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.205004', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f693306-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': '056e1997237611adf164c21c18e693cfc835ee9902d766b8aeea5802dbca1592'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.205004', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f694544-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': 'e4860c842413566919933299d6ac4a61bb3de713c7ed12d894abf812d75bbffe'}]}, 'timestamp': '2025-11-27 09:47:44.206015', '_unique_id': '9ab608c6c9d446b68dec257d274d1bc5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.208 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.208 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9e44f509-f9bf-4352-a95e-02ecc76138f8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.208340', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f69b510-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'b0bfbf67d3b60ebc6ed7bbc0e0286c4e1b579757777abe187a419816f80311ea'}]}, 'timestamp': '2025-11-27 09:47:44.208849', '_unique_id': 'e563343a99e44ca39fbd8ad7932967ad'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.210 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 29130240 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 4300800 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3ec475b3-ef27-4358-bbef-eb27533cba27', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 29130240, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.211059', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6a1d48-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': 'a6fbe71db6c26b7d69fc16d06298ec8022944594ded9e6f86126df7896ec1268'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 4300800, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.211059', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6a2f18-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '808d42981631a73c144cff6fc71d9c36496bbcd6566a672a245b51f138eee65e'}]}, 'timestamp': '2025-11-27 09:47:44.211943', '_unique_id': 'a24b7372066b40ad9b4c88fe6bd494f1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.214 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 498 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '31eaff85-15bf-4d90-acae-a26e7d988b22', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 498, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.214161', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6a967e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '9b4131545c5f22c42c22fd4a589a706edc8425994b4d587283379f46574c1100'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.214161', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6aa826-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '14f322e3e1715b0ddca87338f9a3561b23aa6bd5bb4a946c41bb478edbd9da8f'}]}, 'timestamp': '2025-11-27 09:47:44.215039', '_unique_id': 'b778eb7be8dd4cd7bd2a9645fc2c6ab8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.217 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '86a64fcc-d8f3-4602-99d1-a65437d4a6ee', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.217368', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f6b1536-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': '4e9b06841f814740c67bbcde78e1c2ce95afb1a5a1324d4d5edc77788ce17a07'}]}, 'timestamp': '2025-11-27 09:47:44.217860', '_unique_id': '555a2ab9caf54a71a25d69d47806a8f4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.219 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 95 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4d2728da-1b9d-4356-8d29-15985a1559a9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 95, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.220117', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f6b7f58-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'efe6a1f3851925fa8313860a39ba05bb3009e14e42ffbb4656a636a2d2a3bbbd'}]}, 'timestamp': '2025-11-27 09:47:44.220613', '_unique_id': '8145dec418ae4cb9832e787ac4544610'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.222 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.222 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1064 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 222 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9d8a5c3f-48a0-4bd1-9ee4-f55d449e1366', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1064, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.222796', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6be7a4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '7d146f35906f270d17ecb7e9f7381dc1c6c36b30ebb77908dd042c7d0fcbb14b'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 222, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.222796', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6bf7c6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': 'fa702069efa1e30189f4f62d852bda0c69b39f0016f9d3170fe075c61cd116c4'}]}, 'timestamp': '2025-11-27 09:47:44.223671', '_unique_id': '626f2738efb84cc28beef6a25e894c3e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.225 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.225 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 135 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '85789aa6-cf80-4bdc-83cd-0a301013c9bd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 135, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.225992', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f6c651c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'aedd7f53190c18a69bdf36f61c10f8bb2e1f29c9d71a0c1b96cc0bf3db35db4b'}]}, 'timestamp': '2025-11-27 09:47:44.226477', '_unique_id': '9459cc59b90744c6abb81206c00e4b3d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.228 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.228 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 11748 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '10c7ab32-57f1-4697-8a39-14d817d85b78', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 11748, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.228735', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f6cd150-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': '2e8aed303924202059c672cee855996758c7978a6aa5673e6dda27c58616152f'}]}, 'timestamp': '2025-11-27 09:47:44.229232', '_unique_id': '49ddb0a8d5c34f1e900ce00a6be9ba52'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.231 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.231 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.231 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a859a1a8-24f1-4297-928b-e4e73b9b908a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.231419', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6d3c76-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': '7013b37eda997453356f5e7296850595a0388dc7aac3d40bab4306cb693c063a'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.231419', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6d4e8c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.331855118, 'message_signature': 'f7c0882c586b90c7d610b315b0ab9c6885602068e63fd91c930803f93ef5cb51'}]}, 'timestamp': '2025-11-27 09:47:44.232408', '_unique_id': 'd26dab9c81d14a4f9f48df55966458a1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.234 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 203265086 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 25604091 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1c07c8da-db27-4c76-8f7d-4a675cf801d2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 203265086, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.234416', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6dab02-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '9483fb6163a48dff2119f930494fdede9431e4dff32a425635f8fbf8393f8cc1'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 25604091, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.234416', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6db548-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': 'f42093af330c79d24b2b2febd6009e09320e3dacf407ce9c04b179e6132d3d24'}]}, 'timestamp': '2025-11-27 09:47:44.234960', '_unique_id': '109b61df533e414fa13ba3cfeae72b15'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 73908224 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '93be33bb-ef1b-4ba2-9c5f-435aa2f97e4f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 73908224, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:47:44.236367', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1f6df68e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '52e603291b8f78a15ed22cf34ebdd394466b7d3d3ea945509398aab0eecff326'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:47:44.236367', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1f6e019c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.356932996, 'message_signature': '74b0daebb3eca17cbb040c752b927f3e4ba833914542bc98e7d984917dec4176'}]}, 'timestamp': '2025-11-27 09:47:44.236931', '_unique_id': '1b40bbfd0e8943debee99a4fddc7b24a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.238 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.238 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 54430000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'df0c1bd2-555f-4327-9214-c9256aaf37cc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 54430000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:47:44.238588', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '1f7071de-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.440389731, 'message_signature': 'b1c373e8793e1a24c1dc6f9f77d489b44357029f17603d075f58364ccc26257b'}]}, 'timestamp': '2025-11-27 09:47:44.252980', '_unique_id': '6f82f77a5d0e4db9a4b1d596f4faedcb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.254 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.254 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 434 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b4eb77f5-ed35-41ba-8c9c-2b2c2f45900e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 434, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:47:44.254850', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1f70ca62-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.321485682, 'message_signature': 'b86dc138150752c204d46efa5a20cb1de5cee38316d38362a4dbca4dd7c0bfa7'}]}, 'timestamp': '2025-11-27 09:47:44.255185', '_unique_id': 'c72b8d83512844b88b3f1d601d414652'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 52.46875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1b941046-df82-408d-a748-90b716339abf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 52.46875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:47:44.256544', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '1f710bb2-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11094.440389731, 'message_signature': '5d9cff11eabef2ad29beb88f9fdbb2a70bc693c6416351878d1bf95ddd78f642'}]}, 'timestamp': '2025-11-27 09:47:44.256842', '_unique_id': 'a7c496907d1e4c3eb0735ff476f31f2e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:47:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:47:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 04:47:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24689 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A08CE90000000001030307) Nov 27 04:47:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24690 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A090DB0000000001030307) Nov 27 04:47:46 localhost nova_compute[284026]: 2025-11-27 09:47:46.907 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4996-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40980 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A093DA0000000001030307) Nov 27 04:47:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24691 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A098DB0000000001030307) Nov 27 04:47:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:47:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:47:48 localhost systemd[1]: tmp-crun.1UHIuu.mount: Deactivated successfully. Nov 27 04:47:49 localhost podman[284541]: 2025-11-27 09:47:49.010231168 +0000 UTC m=+0.096276174 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:47:49 localhost podman[284540]: 2025-11-27 09:47:48.990513901 +0000 UTC m=+0.082942074 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:47:49 localhost podman[284541]: 2025-11-27 09:47:49.047040102 +0000 UTC m=+0.133085108 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:47:49 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:47:49 localhost podman[284540]: 2025-11-27 09:47:49.075017436 +0000 UTC m=+0.167445589 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, container_name=ovn_controller, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, config_id=ovn_controller, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:47:49 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:47:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=52478 DF PROTO=TCP SPT=39452 DPT=9102 SEQ=918903966 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A09BDA0000000001030307) Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.564 284030 DEBUG oslo_concurrency.lockutils [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.565 284030 DEBUG oslo_concurrency.lockutils [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.565 284030 DEBUG nova.compute.manager [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.570 284030 DEBUG nova.compute.manager [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 do_stop_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:3338#033[00m Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.575 284030 DEBUG nova.objects.instance [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'flavor' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:47:50 localhost nova_compute[284026]: 2025-11-27 09:47:50.617 284030 DEBUG nova.virt.libvirt.driver [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Shutting down instance from state 1 _clean_shutdown /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4071#033[00m Nov 27 04:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:47:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 7200.1 total, 600.0 interval#012Cumulative writes: 5039 writes, 22K keys, 5039 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5039 writes, 590 syncs, 8.54 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 25 writes, 82 keys, 25 commit groups, 1.0 writes per commit group, ingest: 0.11 MB, 0.00 MB/s#012Interval WAL: 25 writes, 11 syncs, 2.27 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.909 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4997-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.912 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.912 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.912 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.943 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:51 localhost nova_compute[284026]: 2025-11-27 09:47:51.944 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:47:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24692 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A0A89A0000000001030307) Nov 27 04:47:53 localhost kernel: device tapa2718872-3b left promiscuous mode Nov 27 04:47:53 localhost NetworkManager[5971]: [1764236873.0968] device (tapa2718872-3b): state change: disconnected -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Nov 27 04:47:53 localhost ovn_controller[156436]: 2025-11-27T09:47:53Z|00056|binding|INFO|Releasing lport a2718872-3b82-44ca-8d67-a0e516b2708b from this chassis (sb_readonly=0) Nov 27 04:47:53 localhost ovn_controller[156436]: 2025-11-27T09:47:53Z|00057|binding|INFO|Setting lport a2718872-3b82-44ca-8d67-a0e516b2708b down in Southbound Nov 27 04:47:53 localhost ovn_controller[156436]: 2025-11-27T09:47:53Z|00058|binding|INFO|Removing iface tapa2718872-3b ovn-installed in OVS Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.138 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.145 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost systemd[1]: machine-qemu\x2d1\x2dinstance\x2d00000002.scope: Deactivated successfully. Nov 27 04:47:53 localhost systemd[1]: machine-qemu\x2d1\x2dinstance\x2d00000002.scope: Consumed 3min 40.634s CPU time. Nov 27 04:47:53 localhost systemd-machined[86026]: Machine qemu-1-instance-00000002 terminated. Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.156 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:86:b9:b3 192.168.0.212'], port_security=['fa:16:3e:86:b9:b3 192.168.0.212'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '192.168.0.212/24', 'neutron:device_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'neutron:device_owner': 'compute:nova', 'neutron:host_id': 'np0005537446.localdomain', 'neutron:mtu': '', 'neutron:network_name': 'neutron-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'neutron:port_capabilities': '', 'neutron:port_fip': '192.168.122.20', 'neutron:port_name': '', 'neutron:project_id': '02a308de13ca4518beaddd402ba93936', 'neutron:revision_number': '7', 'neutron:security_group_ids': '05e0d7b2-5469-4d02-a67e-d8080e0339fc 869e5a88-209d-40d4-906d-1e63476c9c16', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=f32dcab3-a58b-44c3-83ba-19ec8eecda08, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=a2718872-3b82-44ca-8d67-a0e516b2708b) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.158 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a2718872-3b82-44ca-8d67-a0e516b2708b in datapath 3d77aad9-8375-4ff0-8227-0972f0af9ab9 unbound from our chassis#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.160 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 9deb14fa-14a7-41f5-bf8e-ac8ae4bba815 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.160 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 3d77aad9-8375-4ff0-8227-0972f0af9ab9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.161 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[5d394507-5b97-417e-aae3-99af6f4fc881]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.162 162092 INFO neutron.agent.ovn.metadata.agent [-] Cleaning up ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9 namespace which is not needed anymore#033[00m Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [NOTICE] (269558) : haproxy version is 2.8.14-c23fe91 Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [NOTICE] (269558) : path to executable is /usr/sbin/haproxy Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [WARNING] (269558) : Exiting Master process... Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [WARNING] (269558) : Exiting Master process... Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [ALERT] (269558) : Current worker (269560) exited with code 143 (Terminated) Nov 27 04:47:53 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[269554]: [WARNING] (269558) : All workers exited. Exiting... (0) Nov 27 04:47:53 localhost systemd[1]: libpod-0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014.scope: Deactivated successfully. Nov 27 04:47:53 localhost podman[284615]: 2025-11-27 09:47:53.342849785 +0000 UTC m=+0.070255421 container died 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.348 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.359 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost systemd[1]: tmp-crun.DLPBUe.mount: Deactivated successfully. Nov 27 04:47:53 localhost podman[284615]: 2025-11-27 09:47:53.389365155 +0000 UTC m=+0.116770781 container cleanup 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 04:47:53 localhost podman[284628]: 2025-11-27 09:47:53.429501306 +0000 UTC m=+0.083053528 container cleanup 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:47:53 localhost systemd[1]: libpod-conmon-0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014.scope: Deactivated successfully. Nov 27 04:47:53 localhost podman[284652]: 2025-11-27 09:47:53.477698318 +0000 UTC m=+0.069675796 container remove 0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.482 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[877ef311-3067-461f-8864-edf767dfff54]: (4, ('Thu Nov 27 09:47:53 AM UTC 2025 Stopping container neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9 (0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014)\n0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014\nThu Nov 27 09:47:53 AM UTC 2025 Deleting container neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9 (0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014)\n0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.484 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e2263bed-2088-4bc3-a100-248cc4b5f254]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.485 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap3d77aad9-80, bridge=None, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:47:53 localhost kernel: device tap3d77aad9-80 left promiscuous mode Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.490 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.500 284030 DEBUG nova.compute.manager [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received event network-vif-unplugged-a2718872-3b82-44ca-8d67-a0e516b2708b external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.501 284030 DEBUG oslo_concurrency.lockutils [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.502 284030 DEBUG oslo_concurrency.lockutils [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.502 284030 DEBUG oslo_concurrency.lockutils [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.502 284030 DEBUG nova.compute.manager [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] No waiting events found dispatching network-vif-unplugged-a2718872-3b82-44ca-8d67-a0e516b2708b pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.503 284030 WARNING nova.compute.manager [req-82cfebf5-5b93-4da1-9760-1a777c11425e req-dd4a1625-99e5-4df2-8000-c2377b108d9b c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received unexpected event network-vif-unplugged-a2718872-3b82-44ca-8d67-a0e516b2708b for instance with vm_state active and task_state powering-off.#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.504 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.505 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[55352d98-e9b0-42da-bc5f-912519bf3939]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.519 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4f2e9f11-0aee-476c-87ce-9bb5112c804a]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.521 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[9765a8cf-a062-4e4f-b070-aea867ed98ef]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.533 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[f06b1965-8a1b-4d66-9304-dc47d0a85814]: (4, [{'family': 0, '__align': (), 'ifi_type': 772, 'index': 1, 'flags': 65609, 'change': 0, 'attrs': [['IFLA_IFNAME', 'lo'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UNKNOWN'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 65536], ['IFLA_MIN_MTU', 0], ['IFLA_MAX_MTU', 0], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 1], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 1], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 0], ['IFLA_CARRIER_UP_COUNT', 0], ['IFLA_CARRIER_DOWN_COUNT', 0], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', '00:00:00:00:00:00'], ['IFLA_BROADCAST', '00:00:00:00:00:00'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 1, 'nopolicy': 1, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 662124, 'reachable_time': 34740, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 65536, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 4294967295, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 4294967295, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 0, 'inoctets': 0, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 0, 'outoctets': 0, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 0, 'outmcastpkts': 0, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 0, 'outmcastoctets': 0, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 0, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 0, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1356, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 284676, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.540 162227 DEBUG neutron.privileged.agent.linux.ip_lib [-] Namespace ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9 deleted. remove_netns /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:607#033[00m Nov 27 04:47:53 localhost ovn_metadata_agent[162087]: 2025-11-27 09:47:53.541 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[7552065c-8c51-4dd0-91b8-9d2dd988a587]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.638 284030 INFO nova.virt.libvirt.driver [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Instance shutdown successfully after 3 seconds.#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.644 284030 INFO nova.virt.libvirt.driver [-] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Instance destroyed successfully.#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.645 284030 DEBUG nova.objects.instance [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'numa_topology' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.661 284030 DEBUG nova.compute.manager [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:47:53 localhost nova_compute[284026]: 2025-11-27 09:47:53.743 284030 DEBUG oslo_concurrency.lockutils [None req-ec3eccf5-18f9-40de-904c-8bef35030f1a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 3.178s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:54 localhost systemd[1]: var-lib-containers-storage-overlay-56bac76312ede99d3db3a0fd8d60c45c239702a13f7e02223b811f27ac89ef11-merged.mount: Deactivated successfully. Nov 27 04:47:54 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-0aa0eaa6bc931de8bcad16869ca2ac5b7e90ad514bee70c57386b2a441f4a014-userdata-shm.mount: Deactivated successfully. Nov 27 04:47:54 localhost systemd[1]: run-netns-ovnmeta\x2d3d77aad9\x2d8375\x2d4ff0\x2d8227\x2d0972f0af9ab9.mount: Deactivated successfully. Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.546 284030 DEBUG nova.compute.manager [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.546 284030 DEBUG oslo_concurrency.lockutils [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.547 284030 DEBUG oslo_concurrency.lockutils [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.547 284030 DEBUG oslo_concurrency.lockutils [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.547 284030 DEBUG nova.compute.manager [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] No waiting events found dispatching network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 04:47:55 localhost nova_compute[284026]: 2025-11-27 09:47:55.548 284030 WARNING nova.compute.manager [req-2d6e214f-6149-4df8-ad49-05e2df3dd165 req-a4b7957d-4c4e-44b9-ae37-6ea7c4d67fe1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received unexpected event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b for instance with vm_state stopped and task_state None.#033[00m Nov 27 04:47:55 localhost openstack_network_exporter[244641]: ERROR 09:47:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:47:55 localhost openstack_network_exporter[244641]: ERROR 09:47:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:47:55 localhost openstack_network_exporter[244641]: ERROR 09:47:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:47:55 localhost openstack_network_exporter[244641]: ERROR 09:47:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:47:55 localhost openstack_network_exporter[244641]: Nov 27 04:47:55 localhost openstack_network_exporter[244641]: ERROR 09:47:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:47:55 localhost openstack_network_exporter[244641]: Nov 27 04:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:47:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 7200.2 total, 600.0 interval#012Cumulative writes: 5593 writes, 24K keys, 5593 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5593 writes, 753 syncs, 7.43 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 24 writes, 51 keys, 24 commit groups, 1.0 writes per commit group, ingest: 0.03 MB, 0.00 MB/s#012Interval WAL: 24 writes, 12 syncs, 2.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:47:56 localhost nova_compute[284026]: 2025-11-27 09:47:56.981 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.565 284030 DEBUG nova.compute.manager [None req-95ae440e-119a-4bb0-9a28-8eb0c5766b44 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server [None req-95ae440e-119a-4bb0-9a28-8eb0c5766b44 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Exception during message handling: nova.exception.InstanceInvalidState: Instance a02f7f2f-d2cf-4612-b18a-8be435257201 in power state shutdown. Cannot get_diagnostics while the instance is in this state. Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server Traceback (most recent call last): Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/server.py", line 165, in _process_incoming Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/exception_wrapper.py", line 71, in wrapped Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server _emit_versioned_exception_notification( Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server self.force_reraise() Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server raise self.value Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/exception_wrapper.py", line 63, in wrapped Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 214, in decorated_function Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server compute_utils.add_instance_fault_from_exc(context, Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server self.force_reraise() Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server raise self.value Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 203, in decorated_function Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 6739, in get_instance_diagnostics Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server raise exception.InstanceInvalidState( Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server nova.exception.InstanceInvalidState: Instance a02f7f2f-d2cf-4612-b18a-8be435257201 in power state shutdown. Cannot get_diagnostics while the instance is in this state. Nov 27 04:48:00 localhost nova_compute[284026]: 2025-11-27 09:48:00.595 284030 ERROR oslo_messaging.rpc.server #033[00m Nov 27 04:48:01 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24693 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A0C9DA0000000001030307) Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.038 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4997-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.040 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.040 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5057 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.040 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.042 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:02 localhost nova_compute[284026]: 2025-11-27 09:48:02.042 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:48:02 localhost systemd[1]: tmp-crun.nhNeCi.mount: Deactivated successfully. Nov 27 04:48:02 localhost podman[284678]: 2025-11-27 09:48:02.999104869 +0000 UTC m=+0.092109744 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.build-date=20251125) Nov 27 04:48:03 localhost podman[284678]: 2025-11-27 09:48:03.015002306 +0000 UTC m=+0.108007171 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, container_name=multipathd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 04:48:03 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:48:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:48:04 localhost systemd[1]: tmp-crun.H7IQ5E.mount: Deactivated successfully. Nov 27 04:48:04 localhost podman[284784]: 2025-11-27 09:48:04.369590538 +0000 UTC m=+0.084262709 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:48:04 localhost podman[284784]: 2025-11-27 09:48:04.401111654 +0000 UTC m=+0.115783805 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:48:04 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:48:05 localhost nova_compute[284026]: 2025-11-27 09:48:05.759 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:05 localhost nova_compute[284026]: 2025-11-27 09:48:05.760 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:05 localhost nova_compute[284026]: 2025-11-27 09:48:05.760 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:48:05 localhost nova_compute[284026]: 2025-11-27 09:48:05.760 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.045 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4999-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.091 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.092 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.519 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.520 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.521 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:48:07 localhost nova_compute[284026]: 2025-11-27 09:48:07.521 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:48:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:48:07 localhost podman[284856]: 2025-11-27 09:48:07.875866457 +0000 UTC m=+0.082733519 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:48:07 localhost podman[284856]: 2025-11-27 09:48:07.891018854 +0000 UTC m=+0.097885946 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 04:48:07 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:48:07 localhost podman[284857]: 2025-11-27 09:48:07.984219496 +0000 UTC m=+0.188375777 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, config_id=edpm, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., container_name=openstack_network_exporter, maintainer=Red Hat, Inc., vcs-type=git, release=1755695350, distribution-scope=public, architecture=x86_64, build-date=2025-08-20T13:12:41, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.tags=minimal rhel9) Nov 27 04:48:07 localhost podman[284857]: 2025-11-27 09:48:07.998585683 +0000 UTC m=+0.202741974 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, maintainer=Red Hat, Inc., managed_by=edpm_ansible, version=9.6, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, vendor=Red Hat, Inc., distribution-scope=public, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm) Nov 27 04:48:08 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.362 284030 DEBUG nova.virt.driver [-] Emitting event Stopped> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.363 284030 INFO nova.compute.manager [-] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] VM Stopped (Lifecycle Event)#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.396 284030 DEBUG nova.compute.manager [None req-377b95af-861b-4ac9-a701-e79ced095707 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.401 284030 DEBUG nova.compute.manager [None req-377b95af-861b-4ac9-a701-e79ced095707 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Synchronizing instance power state after lifecycle event "Stopped"; current vm_state: stopped, current task_state: None, current DB power_state: 4, VM power_state: 4 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.720 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.760 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.761 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:48:08 localhost podman[242678]: time="2025-11-27T09:48:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.762 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.762 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.763 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.763 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.764 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.764 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.765 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.765 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:48:08 localhost podman[242678]: @ - - [27/Nov/2025:09:48:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 149986 "" "Go-http-client/1.1" Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.794 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.795 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.795 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.796 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:48:08 localhost nova_compute[284026]: 2025-11-27 09:48:08.796 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:48:08 localhost podman[242678]: @ - - [27/Nov/2025:09:48:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 17724 "" "Go-http-client/1.1" Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.260 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.463s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.337 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.337 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.539 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.541 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=12267MB free_disk=41.837059020996094GB free_vcpus=8 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.542 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.542 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.634 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.635 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.635 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:48:09 localhost nova_compute[284026]: 2025-11-27 09:48:09.680 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:48:10 localhost nova_compute[284026]: 2025-11-27 09:48:10.140 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.460s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:48:10 localhost nova_compute[284026]: 2025-11-27 09:48:10.148 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:48:10 localhost nova_compute[284026]: 2025-11-27 09:48:10.166 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:48:10 localhost nova_compute[284026]: 2025-11-27 09:48:10.195 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:48:10 localhost nova_compute[284026]: 2025-11-27 09:48:10.196 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.654s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:48:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:48:11 localhost podman[284937]: 2025-11-27 09:48:11.988875732 +0000 UTC m=+0.082296057 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:48:12 localhost podman[284937]: 2025-11-27 09:48:12.001941995 +0000 UTC m=+0.095362320 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:48:12 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.093 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4997-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.095 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.095 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.096 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.121 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:12 localhost nova_compute[284026]: 2025-11-27 09:48:12.122 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45741 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A102180000000001030307) Nov 27 04:48:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45742 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1061A0000000001030307) Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.123 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4997-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.125 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.125 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.125 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.148 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:17 localhost nova_compute[284026]: 2025-11-27 09:48:17.149 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24694 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A109DA0000000001030307) Nov 27 04:48:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45743 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A10E1A0000000001030307) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.225 284030 DEBUG nova.compute.manager [None req-8a4f672d-bdec-46a8-8d57-8658e387108a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server [None req-8a4f672d-bdec-46a8-8d57-8658e387108a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Exception during message handling: nova.exception.InstanceInvalidState: Instance a02f7f2f-d2cf-4612-b18a-8be435257201 in power state shutdown. Cannot get_diagnostics while the instance is in this state. Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server Traceback (most recent call last): Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/server.py", line 165, in _process_incoming Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/exception_wrapper.py", line 71, in wrapped Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server _emit_versioned_exception_notification( Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server self.force_reraise() Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server raise self.value Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/exception_wrapper.py", line 63, in wrapped Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 214, in decorated_function Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server compute_utils.add_instance_fault_from_exc(context, Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server self.force_reraise() Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server raise self.value Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 203, in decorated_function Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 6739, in get_instance_diagnostics Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server raise exception.InstanceInvalidState( Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server nova.exception.InstanceInvalidState: Instance a02f7f2f-d2cf-4612-b18a-8be435257201 in power state shutdown. Cannot get_diagnostics while the instance is in this state. Nov 27 04:48:19 localhost nova_compute[284026]: 2025-11-27 09:48:19.254 284030 ERROR oslo_messaging.rpc.server #033[00m Nov 27 04:48:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=40981 DF PROTO=TCP SPT=47254 DPT=9102 SEQ=684201307 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A111DA0000000001030307) Nov 27 04:48:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:48:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:48:19 localhost podman[284958]: 2025-11-27 09:48:19.996677827 +0000 UTC m=+0.088738216 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ovn_controller) Nov 27 04:48:20 localhost systemd[1]: tmp-crun.uEKjxN.mount: Deactivated successfully. Nov 27 04:48:20 localhost podman[284958]: 2025-11-27 09:48:20.051209225 +0000 UTC m=+0.143269604 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible) Nov 27 04:48:20 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:48:20 localhost podman[284959]: 2025-11-27 09:48:20.054909882 +0000 UTC m=+0.142338670 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:48:20 localhost podman[284959]: 2025-11-27 09:48:20.137999649 +0000 UTC m=+0.225428457 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:48:20 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.150 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 4997-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.152 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.152 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: idle 5003 ms, sending inactivity probe run /usr/lib64/python3.9/site-packages/ovs/reconnect.py:117#033[00m Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.153 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering IDLE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.191 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:22 localhost nova_compute[284026]: 2025-11-27 09:48:22.191 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] tcp:127.0.0.1:6640: entering ACTIVE _transition /usr/lib64/python3.9/site-packages/ovs/reconnect.py:519#033[00m Nov 27 04:48:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45744 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A11DDA0000000001030307) Nov 27 04:48:23 localhost ovn_controller[156436]: 2025-11-27T09:48:23Z|00059|memory_trim|INFO|Detected inactivity (last active 30004 ms ago): trimming memory Nov 27 04:48:24 localhost nova_compute[284026]: 2025-11-27 09:48:24.623 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'flavor' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:24 localhost nova_compute[284026]: 2025-11-27 09:48:24.646 284030 DEBUG oslo_concurrency.lockutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:48:24 localhost nova_compute[284026]: 2025-11-27 09:48:24.646 284030 DEBUG oslo_concurrency.lockutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:48:24 localhost nova_compute[284026]: 2025-11-27 09:48:24.647 284030 DEBUG nova.network.neutron [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Building network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2010#033[00m Nov 27 04:48:24 localhost nova_compute[284026]: 2025-11-27 09:48:24.647 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.028 284030 DEBUG nova.network.neutron [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.046 284030 DEBUG oslo_concurrency.lockutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.078 284030 INFO nova.virt.libvirt.driver [-] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Instance destroyed successfully.#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.080 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'numa_topology' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.095 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'resources' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.109 284030 DEBUG nova.virt.libvirt.vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=Flavor(2),hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=,power_state=4,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member,admin',image_base_image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='pc-q35-rhel9.0.0',image_hw_pointer_model='usbtablet',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',image_owner_specified.openstack.md5='',image_owner_specified.openstack.object='images/cirros',image_owner_specified.openstack.sha256='',owner_project_name='admin',owner_user_name='admin'},tags=,task_state='powering-on',terminated_at=None,trusted_certs=,updated_at=2025-11-27T09:47:53Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=,vcpus=1,vm_mode=None,vm_state='stopped') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} unplug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:828#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.110 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.112 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.113 284030 DEBUG os_vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Unplugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') unplug /usr/lib/python3.9/site-packages/os_vif/__init__.py:109#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.117 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.118 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tapa2718872-3b, bridge=br-int, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.159 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.161 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.165 284030 INFO os_vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Successfully unplugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b')#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.169 284030 DEBUG nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Checking UEFI support for host arch (x86_64) supports_uefi /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1754#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.170 284030 INFO nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] UEFI support detected#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.179 284030 DEBUG nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Start _get_guest_xml network_info=[{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] disk_info={'disk_bus': 'virtio', 'cdrom_bus': 'sata', 'mapping': {'root': {'bus': 'virtio', 'dev': 'vda', 'type': 'disk', 'boot_index': '1'}, 'disk': {'bus': 'virtio', 'dev': 'vda', 'type': 'disk', 'boot_index': '1'}, 'disk.eph0': {'bus': 'virtio', 'dev': 'vdb', 'type': 'disk'}}} image_meta=ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='qcow2',id=665842af-e631-4036-a0eb-09cf83c9a80f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=) rescue=None block_device_info={'root_device_name': '/dev/vda', 'image': [{'boot_index': 0, 'device_name': '/dev/vda', 'encrypted': False, 'guest_format': None, 'size': 0, 'encryption_format': None, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': 'virtio', 'encryption_secret_uuid': None, 'image_id': '665842af-e631-4036-a0eb-09cf83c9a80f'}], 'ephemerals': [{'device_name': '/dev/vdb', 'encrypted': False, 'guest_format': None, 'size': 1, 'encryption_format': None, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': 'virtio', 'encryption_secret_uuid': None}], 'block_device_mapping': [], 'swap': None} _get_guest_xml /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:7549#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.185 284030 WARNING nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.187 284030 DEBUG nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Searching host: 'np0005537446.localdomain' for CPU controller through CGroups V1... _has_cgroupsv1_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1653#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.188 284030 DEBUG nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] CPU controller missing on host. _has_cgroupsv1_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1663#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.189 284030 DEBUG nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Searching host: 'np0005537446.localdomain' for CPU controller through CGroups V2... _has_cgroupsv2_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1672#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.190 284030 DEBUG nova.virt.libvirt.host [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] CPU controller found on host. _has_cgroupsv2_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1679#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.190 284030 DEBUG nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] CPU mode 'host-model' models '' was chosen, with extra flags: '' _get_guest_cpu_model_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:5396#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.191 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Getting desirable topologies for flavor Flavor(created_at=2025-11-27T08:32:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={},flavorid='ad8d4a49-18f7-4954-9703-7480b3ae8896',id=2,is_public=True,memory_mb=512,name='m1.small',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='qcow2',id=665842af-e631-4036-a0eb-09cf83c9a80f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: True _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:563#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.191 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Flavor limits 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:348#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.191 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Image limits 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:352#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.192 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Flavor pref 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:388#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.192 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Image pref 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:392#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.192 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:430#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.193 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:569#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.193 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Build topologies for 1 vcpu(s) 1:1:1 _get_possible_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:471#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.193 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Got 1 possible topologies _get_possible_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:501#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.193 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:575#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.194 284030 DEBUG nova.virt.hardware [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:577#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.194 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'vcpu_model' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.237 284030 DEBUG nova.privsep.utils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Path '/var/lib/nova/instances' supports direct I/O supports_direct_io /usr/lib/python3.9/site-packages/nova/privsep/utils.py:63#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.238 284030 DEBUG oslo_concurrency.processutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Running cmd (subprocess): ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.696 284030 DEBUG oslo_concurrency.processutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] CMD "ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.457s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:48:25 localhost nova_compute[284026]: 2025-11-27 09:48:25.698 284030 DEBUG oslo_concurrency.processutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Running cmd (subprocess): ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:48:25 localhost openstack_network_exporter[244641]: ERROR 09:48:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:48:25 localhost openstack_network_exporter[244641]: ERROR 09:48:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:48:25 localhost openstack_network_exporter[244641]: ERROR 09:48:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:48:25 localhost openstack_network_exporter[244641]: ERROR 09:48:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:48:25 localhost openstack_network_exporter[244641]: Nov 27 04:48:25 localhost openstack_network_exporter[244641]: ERROR 09:48:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:48:25 localhost openstack_network_exporter[244641]: Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.202 284030 DEBUG oslo_concurrency.processutils [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] CMD "ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.505s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.205 284030 DEBUG nova.virt.libvirt.vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=Flavor(2),hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=,power_state=4,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member,admin',image_base_image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='pc-q35-rhel9.0.0',image_hw_pointer_model='usbtablet',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',image_owner_specified.openstack.md5='',image_owner_specified.openstack.object='images/cirros',image_owner_specified.openstack.sha256='',owner_project_name='admin',owner_user_name='admin'},tags=,task_state='powering-on',terminated_at=None,trusted_certs=,updated_at=2025-11-27T09:47:53Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=VirtCPUModel,vcpus=1,vm_mode=None,vm_state='stopped') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} virt_type=kvm get_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:563#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.206 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.207 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.210 284030 DEBUG nova.objects.instance [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Lazy-loading 'pci_devices' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.265 284030 DEBUG nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] End _get_guest_xml xml= Nov 27 04:48:26 localhost nova_compute[284026]: a02f7f2f-d2cf-4612-b18a-8be435257201 Nov 27 04:48:26 localhost nova_compute[284026]: instance-00000002 Nov 27 04:48:26 localhost nova_compute[284026]: 524288 Nov 27 04:48:26 localhost nova_compute[284026]: 1 Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: test Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:25 Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: 512 Nov 27 04:48:26 localhost nova_compute[284026]: 1 Nov 27 04:48:26 localhost nova_compute[284026]: 0 Nov 27 04:48:26 localhost nova_compute[284026]: 1 Nov 27 04:48:26 localhost nova_compute[284026]: 1 Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: admin Nov 27 04:48:26 localhost nova_compute[284026]: admin Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: RDO Nov 27 04:48:26 localhost nova_compute[284026]: OpenStack Compute Nov 27 04:48:26 localhost nova_compute[284026]: 27.5.2-0.20250829104910.6f8decf.el9 Nov 27 04:48:26 localhost nova_compute[284026]: a02f7f2f-d2cf-4612-b18a-8be435257201 Nov 27 04:48:26 localhost nova_compute[284026]: a02f7f2f-d2cf-4612-b18a-8be435257201 Nov 27 04:48:26 localhost nova_compute[284026]: Virtual Machine Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: hvm Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: /dev/urandom Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: Nov 27 04:48:26 localhost nova_compute[284026]: _get_guest_xml /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:7555#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.267 284030 DEBUG nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.268 284030 DEBUG nova.virt.libvirt.driver [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.270 284030 DEBUG nova.virt.libvirt.vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T08:33:01Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='test',display_name='test',ec2_ids=,ephemeral_gb=1,ephemeral_key_uuid=None,fault=,flavor=Flavor(2),hidden=False,host='np0005537446.localdomain',hostname='test',id=2,image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',info_cache=InstanceInfoCache,instance_type_id=2,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T08:33:11Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=512,metadata={},migration_context=,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=PciDeviceList,pci_requests=,power_state=4,progress=0,project_id='02a308de13ca4518beaddd402ba93936',ramdisk_id='',reservation_id='r-y06pb0ce',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member,admin',image_base_image_ref='665842af-e631-4036-a0eb-09cf83c9a80f',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='pc-q35-rhel9.0.0',image_hw_pointer_model='usbtablet',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',image_owner_specified.openstack.md5='',image_owner_specified.openstack.object='images/cirros',image_owner_specified.openstack.sha256='',owner_project_name='admin',owner_user_name='admin'},tags=,task_state='powering-on',terminated_at=None,trusted_certs=,updated_at=2025-11-27T09:47:53Z,user_data=None,user_id='3911b9fb06a84c8eb08d5c917916dc45',uuid=a02f7f2f-d2cf-4612-b18a-8be435257201,vcpu_model=VirtCPUModel,vcpus=1,vm_mode=None,vm_state='stopped') vif={"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.270 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converting VIF {"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.271 284030 DEBUG nova.network.os_vif_util [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.272 284030 DEBUG os_vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Plugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.273 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.273 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.274 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.278 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.278 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tapa2718872-3b, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.279 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tapa2718872-3b, col_values=(('external_ids', {'iface-id': 'a2718872-3b82-44ca-8d67-a0e516b2708b', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:86:b9:b3', 'vm-uuid': 'a02f7f2f-d2cf-4612-b18a-8be435257201'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.308 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.311 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.316 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.317 284030 INFO os_vif [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] Successfully plugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:86:b9:b3,bridge_name='br-int',has_traffic_filtering=True,id=a2718872-3b82-44ca-8d67-a0e516b2708b,network=Network(3d77aad9-8375-4ff0-8227-0972f0af9ab9),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapa2718872-3b')#033[00m Nov 27 04:48:26 localhost systemd[1]: Started libvirt secret daemon. Nov 27 04:48:26 localhost kernel: device tapa2718872-3b entered promiscuous mode Nov 27 04:48:26 localhost NetworkManager[5971]: [1764236906.4410] manager: (tapa2718872-3b): new Tun device (/org/freedesktop/NetworkManager/Devices/17) Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.444 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost ovn_controller[156436]: 2025-11-27T09:48:26Z|00060|binding|INFO|Claiming lport a2718872-3b82-44ca-8d67-a0e516b2708b for this chassis. Nov 27 04:48:26 localhost ovn_controller[156436]: 2025-11-27T09:48:26Z|00061|binding|INFO|a2718872-3b82-44ca-8d67-a0e516b2708b: Claiming fa:16:3e:86:b9:b3 192.168.0.212 Nov 27 04:48:26 localhost systemd-udevd[285079]: Network interface NamePolicy= disabled on kernel command line. Nov 27 04:48:26 localhost ovn_controller[156436]: 2025-11-27T09:48:26Z|00062|binding|INFO|Setting lport a2718872-3b82-44ca-8d67-a0e516b2708b ovn-installed in OVS Nov 27 04:48:26 localhost ovn_controller[156436]: 2025-11-27T09:48:26Z|00063|binding|INFO|Setting lport a2718872-3b82-44ca-8d67-a0e516b2708b up in Southbound Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.456 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:86:b9:b3 192.168.0.212'], port_security=['fa:16:3e:86:b9:b3 192.168.0.212'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '192.168.0.212/24', 'neutron:device_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'neutron:device_owner': 'compute:nova', 'neutron:mtu': '', 'neutron:network_name': 'neutron-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'neutron:port_capabilities': '', 'neutron:port_fip': '192.168.122.20', 'neutron:port_name': '', 'neutron:project_id': '02a308de13ca4518beaddd402ba93936', 'neutron:revision_number': '8', 'neutron:security_group_ids': '05e0d7b2-5469-4d02-a67e-d8080e0339fc 869e5a88-209d-40d4-906d-1e63476c9c16', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=f32dcab3-a58b-44c3-83ba-19ec8eecda08, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=a2718872-3b82-44ca-8d67-a0e516b2708b) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.458 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.459 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.460 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a2718872-3b82-44ca-8d67-a0e516b2708b in datapath 3d77aad9-8375-4ff0-8227-0972f0af9ab9 bound to our chassis#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.463 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 9deb14fa-14a7-41f5-bf8e-ac8ae4bba815 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.463 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network 3d77aad9-8375-4ff0-8227-0972f0af9ab9#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.476 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[bd3058c5-4dc6-480b-a8d4-e8ddf538ee9b]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.478 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Creating VETH tap3d77aad9-81 in ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9 namespace provision_datapath /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:665#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.481 162196 DEBUG neutron.privileged.agent.linux.ip_lib [-] Interface tap3d77aad9-80 not found in namespace None get_link_id /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:204#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.481 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[49d2cad0-e3fb-459a-9f61-92f630493cc7]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.482 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4419b8fe-0c4a-42ed-8c05-249bcf11df79]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost NetworkManager[5971]: [1764236906.4889] device (tapa2718872-3b): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 04:48:26 localhost NetworkManager[5971]: [1764236906.4896] device (tapa2718872-3b): state change: unavailable -> disconnected (reason 'none', sys-iface-state: 'external') Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.490 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.496 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.510 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[d482191e-b4a2-45ba-9f1b-d8983f754dfb]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost systemd-machined[86026]: New machine qemu-2-instance-00000002. Nov 27 04:48:26 localhost systemd[1]: Started Virtual Machine qemu-2-instance-00000002. Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.543 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[7ad88273-9c38-4072-b5d6-90dc68792f14]: (4, ('net.ipv4.conf.all.promote_secondaries = 1\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.580 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[0769445b-8785-4b05-920f-be4ce138bdd4]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.589 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8917e29a-2a3b-4dd7-a299-6f503c5d863b]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost NetworkManager[5971]: [1764236906.5911] manager: (tap3d77aad9-80): new Veth device (/org/freedesktop/NetworkManager/Devices/18) Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.621 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[f243d6a5-ab45-408a-a2d2-29b8a15feac3]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.628 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[5fb5268a-a63b-4020-93a9-685dabd8e0be]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap3d77aad9-81: link becomes ready Nov 27 04:48:26 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap3d77aad9-80: link becomes ready Nov 27 04:48:26 localhost NetworkManager[5971]: [1764236906.6513] device (tap3d77aad9-80): carrier: link connected Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.663 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[f0a4450c-0a14-48e7-b592-23ff4318d277]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.681 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c87ec918-8502-46c7-b331-6798c6c18e18]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap3d77aad9-81'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:76:bb:fc'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 19], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1113676, 'reachable_time': 22324, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 285130, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.699 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b2fdd527-bbdf-47f8-844a-6ea74eba39f1]: (4, ({'family': 10, 'prefixlen': 64, 'flags': 192, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:fe76:bbfc'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 1113676, 'tstamp': 1113676}], ['IFA_FLAGS', 192]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 285133, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'},)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.720 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[7cadc7e9-da52-48bd-bdf8-8d6739e31351]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap3d77aad9-81'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:76:bb:fc'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 19], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1113676, 'reachable_time': 22324, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 0, 'sequence_number': 255, 'pid': 285136, 'error': None, 'target': 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.753 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b2aeb4e0-a5df-4089-a1dc-c33baeb3dec7]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.812 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0d47866a-d614-4de8-9df4-0af20289637a]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.813 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap3d77aad9-80, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.814 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.815 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap3d77aad9-80, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost kernel: device tap3d77aad9-80 entered promiscuous mode Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.817 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.820 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.821 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tap3d77aad9-80, col_values=(('external_ids', {'iface-id': '60750ee5-275f-4987-8dcc-15b61f3cb86a'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 04:48:26 localhost ovn_controller[156436]: 2025-11-27T09:48:26Z|00064|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.826 284030 DEBUG nova.compute.manager [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.826 284030 DEBUG oslo_concurrency.lockutils [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.827 284030 DEBUG oslo_concurrency.lockutils [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.827 284030 DEBUG oslo_concurrency.lockutils [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.828 284030 DEBUG nova.compute.manager [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] No waiting events found dispatching network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.828 284030 WARNING nova.compute.manager [req-238a6608-483f-4561-a87b-5ffe8e48c4a2 req-a7da83d4-a7a3-4852-acf8-1e3d3aa74911 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received unexpected event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b for instance with vm_state stopped and task_state powering-on.#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.828 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.832 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.834 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.836 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[19f9222d-3986-4f8b-8b4f-6a5dc490eead]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.838 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: global Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9 Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: user root Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: group root Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/3d77aad9-8375-4ff0-8227-0972f0af9ab9.pid.haproxy Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: daemon Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: defaults Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: log global Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: mode http Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: option httplog Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: listen listener Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID 3d77aad9-8375-4ff0-8227-0972f0af9ab9 Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 04:48:26 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:26.839 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'env', 'PROCESS_TAG=haproxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/3d77aad9-8375-4ff0-8227-0972f0af9ab9.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.910 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Resumed> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.911 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] VM Resumed (Lifecycle Event)#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.933 284030 DEBUG nova.compute.manager [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Instance event wait completed in 0 seconds for wait_for_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:577#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.937 284030 INFO nova.virt.libvirt.driver [-] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Instance rebooted successfully.#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.938 284030 DEBUG nova.compute.manager [None req-cdac4211-749a-41ed-9577-bf98dea05098 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.941 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:26 localhost nova_compute[284026]: 2025-11-27 09:48:26.953 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Synchronizing instance power state after lifecycle event "Resumed"; current vm_state: stopped, current task_state: powering-on, current DB power_state: 4, VM power_state: 1 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 04:48:27 localhost snmpd[71019]: IfIndex of an interface changed. Such interfaces will appear multiple times in IF-MIB. Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.067 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] During sync_power_state the instance has a pending task (powering-on). Skip.#033[00m Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.068 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Started> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.068 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] VM Started (Lifecycle Event)#033[00m Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.182 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.186 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Synchronizing instance power state after lifecycle event "Started"; current vm_state: active, current task_state: None, current DB power_state: 1, VM power_state: 1 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 04:48:27 localhost nova_compute[284026]: 2025-11-27 09:48:27.195 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:27 localhost podman[285193]: Nov 27 04:48:27 localhost podman[285193]: 2025-11-27 09:48:27.28423467 +0000 UTC m=+0.104417856 container create 83787e16bc900e6cff07cd5145e1c2dd48f420e7ce035952a7a5a7ea28f3b8cb (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 04:48:27 localhost podman[285193]: 2025-11-27 09:48:27.231793646 +0000 UTC m=+0.051976842 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 04:48:27 localhost systemd[1]: Started libpod-conmon-83787e16bc900e6cff07cd5145e1c2dd48f420e7ce035952a7a5a7ea28f3b8cb.scope. Nov 27 04:48:27 localhost systemd[1]: Started libcrun container. Nov 27 04:48:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/a34e7128165f874ab2570298d6322c0a683f6e63284e02a4d4018cfe85b5efb1/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 04:48:27 localhost podman[285193]: 2025-11-27 09:48:27.365318475 +0000 UTC m=+0.185501661 container init 83787e16bc900e6cff07cd5145e1c2dd48f420e7ce035952a7a5a7ea28f3b8cb (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 04:48:27 localhost podman[285193]: 2025-11-27 09:48:27.373673234 +0000 UTC m=+0.193856430 container start 83787e16bc900e6cff07cd5145e1c2dd48f420e7ce035952a7a5a7ea28f3b8cb (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:48:27 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285207]: [NOTICE] (285211) : New worker (285213) forked Nov 27 04:48:27 localhost neutron-haproxy-ovnmeta-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285207]: [NOTICE] (285211) : Loading success. Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.963 284030 DEBUG nova.compute.manager [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.963 284030 DEBUG oslo_concurrency.lockutils [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.964 284030 DEBUG oslo_concurrency.lockutils [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.964 284030 DEBUG oslo_concurrency.lockutils [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.965 284030 DEBUG nova.compute.manager [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] No waiting events found dispatching network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 04:48:28 localhost nova_compute[284026]: 2025-11-27 09:48:28.965 284030 WARNING nova.compute.manager [req-d5d3404d-513a-4994-8962-168afeccf6f1 req-c4d33272-3ecf-4438-be4a-86183f2de2c1 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Received unexpected event network-vif-plugged-a2718872-3b82-44ca-8d67-a0e516b2708b for instance with vm_state active and task_state None.#033[00m Nov 27 04:48:30 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45745 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A13DDA0000000001030307) Nov 27 04:48:31 localhost nova_compute[284026]: 2025-11-27 09:48:31.308 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:32 localhost nova_compute[284026]: 2025-11-27 09:48:32.244 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:48:34 localhost podman[285222]: 2025-11-27 09:48:34.00600007 +0000 UTC m=+0.100862924 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3) Nov 27 04:48:34 localhost podman[285222]: 2025-11-27 09:48:34.022579173 +0000 UTC m=+0.117442027 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:48:34 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:48:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:48:34 localhost systemd[1]: tmp-crun.TphlCG.mount: Deactivated successfully. Nov 27 04:48:35 localhost podman[285242]: 2025-11-27 09:48:34.999259524 +0000 UTC m=+0.092866344 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent) Nov 27 04:48:35 localhost podman[285242]: 2025-11-27 09:48:35.03310451 +0000 UTC m=+0.126711290 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:48:35 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:48:36 localhost nova_compute[284026]: 2025-11-27 09:48:36.311 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:37 localhost nova_compute[284026]: 2025-11-27 09:48:37.248 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:38 localhost podman[242678]: time="2025-11-27T09:48:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:48:38 localhost podman[242678]: @ - - [27/Nov/2025:09:48:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:48:38 localhost podman[242678]: @ - - [27/Nov/2025:09:48:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18204 "" "Go-http-client/1.1" Nov 27 04:48:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:48:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:48:38 localhost systemd[1]: tmp-crun.wCRFgq.mount: Deactivated successfully. Nov 27 04:48:38 localhost podman[285258]: 2025-11-27 09:48:38.991107215 +0000 UTC m=+0.087616907 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, config_id=edpm, io.buildah.version=1.41.3, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 04:48:39 localhost podman[285258]: 2025-11-27 09:48:39.032933331 +0000 UTC m=+0.129443063 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_id=edpm, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 04:48:39 localhost systemd[1]: tmp-crun.OCWO0a.mount: Deactivated successfully. Nov 27 04:48:39 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:48:39 localhost podman[285259]: 2025-11-27 09:48:39.054302541 +0000 UTC m=+0.148239235 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, maintainer=Red Hat, Inc., release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, build-date=2025-08-20T13:12:41, vcs-type=git, config_id=edpm, com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., distribution-scope=public, architecture=x86_64, container_name=openstack_network_exporter, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 04:48:39 localhost podman[285259]: 2025-11-27 09:48:39.065990377 +0000 UTC m=+0.159927031 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, version=9.6, io.openshift.tags=minimal rhel9, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, release=1755695350, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., container_name=openstack_network_exporter, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, com.redhat.component=ubi9-minimal-container, vendor=Red Hat, Inc., managed_by=edpm_ansible, distribution-scope=public) Nov 27 04:48:39 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:48:39 localhost ovn_controller[156436]: 2025-11-27T09:48:39Z|00004|pinctrl(ovn_pinctrl0)|INFO|DHCPACK fa:16:3e:86:b9:b3 192.168.0.212 Nov 27 04:48:41 localhost nova_compute[284026]: 2025-11-27 09:48:41.313 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:42 localhost nova_compute[284026]: 2025-11-27 09:48:42.251 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:48:42 localhost systemd[1]: tmp-crun.1hXvR0.mount: Deactivated successfully. Nov 27 04:48:42 localhost podman[285294]: 2025-11-27 09:48:42.984618361 +0000 UTC m=+0.084317011 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:48:42 localhost podman[285294]: 2025-11-27 09:48:42.998943686 +0000 UTC m=+0.098642366 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:48:43 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:48:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:43.557 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:48:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:43.558 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:48:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:43.558 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:48:44 localhost snmpd[71019]: empty variable list in _query Nov 27 04:48:44 localhost snmpd[71019]: empty variable list in _query Nov 27 04:48:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25572 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A177490000000001030307) Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:45.645 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:45.648 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/instance-id HTTP/1.0#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:45 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:46 localhost nova_compute[284026]: 2025-11-27 09:48:46.315 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25573 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A17B5B0000000001030307) Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.838 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.839 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/instance-id HTTP/1.1" status: 200 len: 146 time: 1.1908231#033[00m Nov 27 04:48:46 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52464 [27/Nov/2025:09:48:45.644] listener listener/metadata 0/0/0/1194/1194 200 130 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-id HTTP/1.1" Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.854 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.855 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/public-keys HTTP/1.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.890 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/public-keys HTTP/1.1" status: 404 len: 297 time: 0.0354421#033[00m Nov 27 04:48:46 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52478 [27/Nov/2025:09:48:46.853] listener listener/metadata 0/0/0/36/36 404 281 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-keys HTTP/1.1" Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.905 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.906 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/instance-id HTTP/1.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:46 localhost nova_compute[284026]: 2025-11-27 09:48:46.914 284030 DEBUG nova.compute.manager [None req-3fffc075-51be-41dc-ba58-c1eaa47a889a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 04:48:46 localhost nova_compute[284026]: 2025-11-27 09:48:46.920 284030 INFO nova.compute.manager [None req-3fffc075-51be-41dc-ba58-c1eaa47a889a 3911b9fb06a84c8eb08d5c917916dc45 02a308de13ca4518beaddd402ba93936 - - default default] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Retrieving diagnostics#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.921 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:46 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52486 [27/Nov/2025:09:48:46.905] listener listener/metadata 0/0/0/16/16 200 130 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-id HTTP/1.1" Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.922 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/instance-id HTTP/1.1" status: 200 len: 146 time: 0.0158503#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.929 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.930 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/ami-launch-index HTTP/1.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.943 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.943 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/ami-launch-index HTTP/1.1" status: 200 len: 136 time: 0.0136917#033[00m Nov 27 04:48:46 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52500 [27/Nov/2025:09:48:46.928] listener listener/metadata 0/0/0/14/14 200 120 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/ami-launch-index HTTP/1.1" Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.950 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.951 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/instance-type HTTP/1.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.992 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.992 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/instance-type HTTP/1.1" status: 200 len: 143 time: 0.0415335#033[00m Nov 27 04:48:46 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52506 [27/Nov/2025:09:48:46.949] listener listener/metadata 0/0/0/42/42 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/instance-type HTTP/1.1" Nov 27 04:48:46 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.999 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:46.999 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/local-ipv4 HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.015 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52516 [27/Nov/2025:09:48:46.998] listener listener/metadata 0/0/0/17/17 200 133 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/local-ipv4 HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.016 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/local-ipv4 HTTP/1.1" status: 200 len: 149 time: 0.0164258#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.022 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.023 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/public-ipv4 HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.038 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52522 [27/Nov/2025:09:48:47.022] listener listener/metadata 0/0/0/16/16 200 134 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-ipv4 HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.038 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/public-ipv4 HTTP/1.1" status: 200 len: 150 time: 0.0151758#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.045 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.045 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/hostname HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.061 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52526 [27/Nov/2025:09:48:47.044] listener listener/metadata 0/0/0/16/16 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/hostname HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.061 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/hostname HTTP/1.1" status: 200 len: 139 time: 0.0159042#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.068 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.068 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/local-hostname HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.081 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52538 [27/Nov/2025:09:48:47.067] listener listener/metadata 0/0/0/13/13 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/local-hostname HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.081 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/local-hostname HTTP/1.1" status: 200 len: 139 time: 0.0129809#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.088 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.089 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/user-data HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52548 [27/Nov/2025:09:48:47.087] listener listener/metadata 0/0/0/17/17 404 281 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/user-data HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.104 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/user-data HTTP/1.1" status: 404 len: 297 time: 0.0159364#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.119 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.120 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/block-device-mapping HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.133 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52552 [27/Nov/2025:09:48:47.119] listener listener/metadata 0/0/0/15/15 200 139 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.134 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/block-device-mapping HTTP/1.1" status: 200 len: 155 time: 0.0144677#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.139 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.140 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/block-device-mapping/ami HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.154 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52564 [27/Nov/2025:09:48:47.138] listener listener/metadata 0/0/0/16/16 200 122 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/ami HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.155 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/block-device-mapping/ami HTTP/1.1" status: 200 len: 138 time: 0.0151718#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.160 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.161 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/block-device-mapping/ephemeral0 HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.178 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52574 [27/Nov/2025:09:48:47.159] listener listener/metadata 0/0/0/19/19 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/ephemeral0 HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.179 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/block-device-mapping/ephemeral0 HTTP/1.1" status: 200 len: 143 time: 0.0183961#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.184 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.185 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/block-device-mapping/root HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.199 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52588 [27/Nov/2025:09:48:47.184] listener listener/metadata 0/0/0/15/15 200 127 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/block-device-mapping/root HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.199 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/block-device-mapping/root HTTP/1.1" status: 200 len: 143 time: 0.0145974#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.206 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.206 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/public-hostname HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.221 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52598 [27/Nov/2025:09:48:47.205] listener listener/metadata 0/0/0/16/16 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/public-hostname HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.222 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/public-hostname HTTP/1.1" status: 200 len: 139 time: 0.0153785#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.229 162191 DEBUG eventlet.wsgi.server [-] (162191) accepted '' server /usr/lib/python3.9/site-packages/eventlet/wsgi.py:1004#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.229 162191 DEBUG neutron.agent.ovn.metadata.server [-] Request: GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Accept: */*#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Connection: close#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Content-Type: text/plain#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: Host: 169.254.169.254#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: User-Agent: curl/7.84.0#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Forwarded-For: 192.168.0.212#015 Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: X-Ovn-Network-Id: 3d77aad9-8375-4ff0-8227-0972f0af9ab9 __call__ /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:82#033[00m Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.242 162191 DEBUG neutron.agent.ovn.metadata.server [-] _proxy_request /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/server.py:161#033[00m Nov 27 04:48:47 localhost haproxy-metadata-proxy-3d77aad9-8375-4ff0-8227-0972f0af9ab9[285213]: 192.168.0.212:52608 [27/Nov/2025:09:48:47.228] listener listener/metadata 0/0/0/14/14 200 123 - - ---- 1/1/0/0/0 0/0 "GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.1" Nov 27 04:48:47 localhost ovn_metadata_agent[162087]: 2025-11-27 09:48:47.243 162191 INFO eventlet.wsgi.server [-] 192.168.0.212, "GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.1" status: 200 len: 139 time: 0.0134513#033[00m Nov 27 04:48:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45746 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A17DDB0000000001030307) Nov 27 04:48:47 localhost nova_compute[284026]: 2025-11-27 09:48:47.277 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25574 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1835A0000000001030307) Nov 27 04:48:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=24695 DF PROTO=TCP SPT=38442 DPT=9102 SEQ=941299197 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A187DA0000000001030307) Nov 27 04:48:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:48:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:48:51 localhost podman[285318]: 2025-11-27 09:48:51.006903595 +0000 UTC m=+0.096102459 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 04:48:51 localhost systemd[1]: tmp-crun.3lt3nX.mount: Deactivated successfully. Nov 27 04:48:51 localhost podman[285318]: 2025-11-27 09:48:51.059668198 +0000 UTC m=+0.148867022 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, org.label-schema.vendor=CentOS, container_name=ovn_controller, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 04:48:51 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:48:51 localhost podman[285319]: 2025-11-27 09:48:51.059020461 +0000 UTC m=+0.148208074 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:48:51 localhost podman[285319]: 2025-11-27 09:48:51.141017769 +0000 UTC m=+0.230205342 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:48:51 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:48:51 localhost nova_compute[284026]: 2025-11-27 09:48:51.317 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:52 localhost nova_compute[284026]: 2025-11-27 09:48:52.279 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25575 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1931A0000000001030307) Nov 27 04:48:55 localhost openstack_network_exporter[244641]: ERROR 09:48:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:48:55 localhost openstack_network_exporter[244641]: ERROR 09:48:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:48:55 localhost openstack_network_exporter[244641]: ERROR 09:48:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:48:55 localhost openstack_network_exporter[244641]: ERROR 09:48:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:48:55 localhost openstack_network_exporter[244641]: Nov 27 04:48:55 localhost openstack_network_exporter[244641]: ERROR 09:48:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:48:55 localhost openstack_network_exporter[244641]: Nov 27 04:48:56 localhost nova_compute[284026]: 2025-11-27 09:48:56.319 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:48:56 localhost ovn_controller[156436]: 2025-11-27T09:48:56Z|00065|memory_trim|INFO|Detected inactivity (last active 30001 ms ago): trimming memory Nov 27 04:48:57 localhost nova_compute[284026]: 2025-11-27 09:48:57.282 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:01 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25576 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1B3DB0000000001030307) Nov 27 04:49:01 localhost nova_compute[284026]: 2025-11-27 09:49:01.321 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:02 localhost nova_compute[284026]: 2025-11-27 09:49:02.285 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:49:04 localhost podman[285365]: 2025-11-27 09:49:04.983005196 +0000 UTC m=+0.077088060 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:49:04 localhost podman[285365]: 2025-11-27 09:49:04.997949878 +0000 UTC m=+0.092032762 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 04:49:05 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:49:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:49:05 localhost podman[285384]: 2025-11-27 09:49:05.984420495 +0000 UTC m=+0.084056234 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent) Nov 27 04:49:06 localhost podman[285384]: 2025-11-27 09:49:06.014942724 +0000 UTC m=+0.114578473 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:49:06 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:49:06 localhost nova_compute[284026]: 2025-11-27 09:49:06.323 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:07 localhost nova_compute[284026]: 2025-11-27 09:49:07.289 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:08 localhost podman[242678]: time="2025-11-27T09:49:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:49:08 localhost podman[242678]: @ - - [27/Nov/2025:09:49:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:49:08 localhost podman[242678]: @ - - [27/Nov/2025:09:49:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18206 "" "Go-http-client/1.1" Nov 27 04:49:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:49:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:49:09 localhost systemd[1]: tmp-crun.evYaBP.mount: Deactivated successfully. Nov 27 04:49:09 localhost podman[285506]: 2025-11-27 09:49:09.231848622 +0000 UTC m=+0.094898508 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:49:09 localhost podman[285506]: 2025-11-27 09:49:09.247866342 +0000 UTC m=+0.110916218 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, tcib_managed=true, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 04:49:09 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:49:09 localhost podman[285507]: 2025-11-27 09:49:09.329635934 +0000 UTC m=+0.189009653 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.openshift.tags=minimal rhel9, vcs-type=git, version=9.6, release=1755695350, name=ubi9-minimal, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., io.buildah.version=1.33.7, architecture=x86_64, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., config_id=edpm) Nov 27 04:49:09 localhost podman[285507]: 2025-11-27 09:49:09.372006134 +0000 UTC m=+0.231379873 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, vcs-type=git, version=9.6, name=ubi9-minimal, vendor=Red Hat, Inc., config_id=edpm, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, maintainer=Red Hat, Inc., build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, release=1755695350, container_name=openstack_network_exporter, io.buildah.version=1.33.7, managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}) Nov 27 04:49:09 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:49:09 localhost podman[285586]: Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.743761955 +0000 UTC m=+0.079349861 container create 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, io.buildah.version=1.33.12, vcs-type=git, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, version=7, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, io.openshift.tags=rhceph ceph, ceph=True, vendor=Red Hat, Inc., name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main) Nov 27 04:49:09 localhost systemd[1]: Started libpod-conmon-51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf.scope. Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.710710399 +0000 UTC m=+0.046298315 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:49:09 localhost systemd[1]: Started libcrun container. Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.832425818 +0000 UTC m=+0.168013714 container init 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, version=7, RELEASE=main, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, GIT_CLEAN=True, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, distribution-scope=public, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=) Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.843308723 +0000 UTC m=+0.178896609 container start 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, maintainer=Guillaume Abrioux , ceph=True, version=7, distribution-scope=public, vendor=Red Hat, Inc., release=553, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, RELEASE=main) Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.843615231 +0000 UTC m=+0.179203157 container attach 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, release=553, maintainer=Guillaume Abrioux , GIT_BRANCH=main, version=7, description=Red Hat Ceph Storage 7, RELEASE=main, vendor=Red Hat, Inc., name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, vcs-type=git, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, ceph=True) Nov 27 04:49:09 localhost cranky_bohr[285602]: 167 167 Nov 27 04:49:09 localhost systemd[1]: libpod-51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf.scope: Deactivated successfully. Nov 27 04:49:09 localhost podman[285586]: 2025-11-27 09:49:09.84891144 +0000 UTC m=+0.184499356 container died 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, description=Red Hat Ceph Storage 7, distribution-scope=public, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, RELEASE=main, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, ceph=True, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, GIT_BRANCH=main, release=553) Nov 27 04:49:09 localhost podman[285607]: 2025-11-27 09:49:09.946801035 +0000 UTC m=+0.084660319 container remove 51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_bohr, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, distribution-scope=public, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, version=7, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, RELEASE=main, architecture=x86_64, CEPH_POINT_RELEASE=, io.openshift.expose-services=, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, vendor=Red Hat, Inc., GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux ) Nov 27 04:49:09 localhost systemd[1]: libpod-conmon-51783c268d3553326ed2e022cc98ee8094601105cbd3bd94b121b3a882d82ccf.scope: Deactivated successfully. Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.164 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.165 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost podman[285629]: Nov 27 04:49:10 localhost podman[285629]: 2025-11-27 09:49:10.178984388 +0000 UTC m=+0.079817492 container create 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, release=553, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, CEPH_POINT_RELEASE=, RELEASE=main, build-date=2025-09-24T08:57:55, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_BRANCH=main, distribution-scope=public, io.buildah.version=1.33.12, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.189 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.189 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.189 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:49:10 localhost systemd[1]: var-lib-containers-storage-overlay-e689ddc239e9213d35e46e3f729d90589b52aee2b55d30c878dda8e011db3f4a-merged.mount: Deactivated successfully. Nov 27 04:49:10 localhost systemd[1]: Started libpod-conmon-7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d.scope. Nov 27 04:49:10 localhost systemd[1]: Started libcrun container. Nov 27 04:49:10 localhost podman[285629]: 2025-11-27 09:49:10.147242327 +0000 UTC m=+0.048075451 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:49:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5303329f8360b18eb812415359c56ead20a260b5bb37f200983d73e85e9a6893/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 04:49:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5303329f8360b18eb812415359c56ead20a260b5bb37f200983d73e85e9a6893/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:49:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5303329f8360b18eb812415359c56ead20a260b5bb37f200983d73e85e9a6893/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 04:49:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5303329f8360b18eb812415359c56ead20a260b5bb37f200983d73e85e9a6893/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:49:10 localhost podman[285629]: 2025-11-27 09:49:10.259779205 +0000 UTC m=+0.160612309 container init 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, ceph=True, release=553, io.buildah.version=1.33.12, vcs-type=git, RELEASE=main, name=rhceph, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, architecture=x86_64, vendor=Red Hat, Inc., GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:49:10 localhost podman[285629]: 2025-11-27 09:49:10.271280386 +0000 UTC m=+0.172113490 container start 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, release=553, io.buildah.version=1.33.12, RELEASE=main, vcs-type=git, architecture=x86_64, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , GIT_CLEAN=True, io.openshift.expose-services=, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, CEPH_POINT_RELEASE=, version=7, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:49:10 localhost podman[285629]: 2025-11-27 09:49:10.271566404 +0000 UTC m=+0.172399548 container attach 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, build-date=2025-09-24T08:57:55, vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , distribution-scope=public, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, architecture=x86_64, release=553, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, GIT_CLEAN=True, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, ceph=True) Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.586 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.587 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.587 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.587 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.962 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.989 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.989 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.990 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.990 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.991 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.991 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.991 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.991 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.992 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:49:10 localhost nova_compute[284026]: 2025-11-27 09:49:10.992 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.015 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.015 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.016 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.016 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.016 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.325 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:11 localhost exciting_jepsen[285644]: [ Nov 27 04:49:11 localhost exciting_jepsen[285644]: { Nov 27 04:49:11 localhost exciting_jepsen[285644]: "available": false, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "ceph_device": false, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "lsm_data": {}, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "lvs": [], Nov 27 04:49:11 localhost exciting_jepsen[285644]: "path": "/dev/sr0", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "rejected_reasons": [ Nov 27 04:49:11 localhost exciting_jepsen[285644]: "Has a FileSystem", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "Insufficient space (<5GB)" Nov 27 04:49:11 localhost exciting_jepsen[285644]: ], Nov 27 04:49:11 localhost exciting_jepsen[285644]: "sys_api": { Nov 27 04:49:11 localhost exciting_jepsen[285644]: "actuators": null, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "device_nodes": "sr0", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "human_readable_size": "482.00 KB", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "id_bus": "ata", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "model": "QEMU DVD-ROM", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "nr_requests": "2", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "partitions": {}, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "path": "/dev/sr0", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "removable": "1", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "rev": "2.5+", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "ro": "0", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "rotational": "1", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "sas_address": "", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "sas_device_handle": "", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "scheduler_mode": "mq-deadline", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "sectors": 0, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "sectorsize": "2048", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "size": 493568.0, Nov 27 04:49:11 localhost exciting_jepsen[285644]: "support_discard": "0", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "type": "disk", Nov 27 04:49:11 localhost exciting_jepsen[285644]: "vendor": "QEMU" Nov 27 04:49:11 localhost exciting_jepsen[285644]: } Nov 27 04:49:11 localhost exciting_jepsen[285644]: } Nov 27 04:49:11 localhost exciting_jepsen[285644]: ] Nov 27 04:49:11 localhost systemd[1]: libpod-7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d.scope: Deactivated successfully. Nov 27 04:49:11 localhost systemd[1]: libpod-7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d.scope: Consumed 1.088s CPU time. Nov 27 04:49:11 localhost podman[287610]: 2025-11-27 09:49:11.439725221 +0000 UTC m=+0.071024431 container died 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, RELEASE=main, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, version=7, ceph=True, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, name=rhceph, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, vcs-type=git, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:49:11 localhost systemd[1]: tmp-crun.IGFmQr.mount: Deactivated successfully. Nov 27 04:49:11 localhost systemd[1]: var-lib-containers-storage-overlay-5303329f8360b18eb812415359c56ead20a260b5bb37f200983d73e85e9a6893-merged.mount: Deactivated successfully. Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.477 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.461s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:49:11 localhost podman[287610]: 2025-11-27 09:49:11.488977102 +0000 UTC m=+0.120276292 container remove 7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_jepsen, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, GIT_BRANCH=main, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, release=553, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , version=7, distribution-scope=public, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, ceph=True) Nov 27 04:49:11 localhost systemd[1]: libpod-conmon-7ddfa0c5574d8b4ee37a812fd3ad881f6bc605cd012cc512e6456fe45ee30b7d.scope: Deactivated successfully. Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.549 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.549 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.735 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.736 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11917MB free_disk=41.83700180053711GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.736 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.736 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:49:11 localhost nova_compute[284026]: 2025-11-27 09:49:11.873 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.278 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.405s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.286 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.294 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.303 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.330 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:49:12 localhost nova_compute[284026]: 2025-11-27 09:49:12.330 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.594s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:49:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:49:13 localhost podman[287667]: 2025-11-27 09:49:13.998852393 +0000 UTC m=+0.087612307 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:49:14 localhost podman[287667]: 2025-11-27 09:49:14.009946244 +0000 UTC m=+0.098706158 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:49:14 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:49:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48515 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1EC790000000001030307) Nov 27 04:49:16 localhost nova_compute[284026]: 2025-11-27 09:49:16.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48516 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1F09A0000000001030307) Nov 27 04:49:17 localhost nova_compute[284026]: 2025-11-27 09:49:17.296 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25577 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1F3DA0000000001030307) Nov 27 04:49:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48517 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1F89B0000000001030307) Nov 27 04:49:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=45747 DF PROTO=TCP SPT=52782 DPT=9102 SEQ=920693155 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A1FBDB0000000001030307) Nov 27 04:49:21 localhost nova_compute[284026]: 2025-11-27 09:49:21.331 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:49:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:49:22 localhost podman[287691]: 2025-11-27 09:49:21.997553908 +0000 UTC m=+0.084812383 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:49:22 localhost podman[287691]: 2025-11-27 09:49:22.009933252 +0000 UTC m=+0.097191747 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:49:22 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:49:22 localhost podman[287690]: 2025-11-27 09:49:22.20031036 +0000 UTC m=+0.291677683 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_id=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 04:49:22 localhost podman[287690]: 2025-11-27 09:49:22.240967375 +0000 UTC m=+0.332334688 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:49:22 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:49:22 localhost nova_compute[284026]: 2025-11-27 09:49:22.297 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48518 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2085A0000000001030307) Nov 27 04:49:25 localhost openstack_network_exporter[244641]: ERROR 09:49:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:49:25 localhost openstack_network_exporter[244641]: ERROR 09:49:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:49:25 localhost openstack_network_exporter[244641]: ERROR 09:49:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:49:25 localhost openstack_network_exporter[244641]: ERROR 09:49:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:49:25 localhost openstack_network_exporter[244641]: Nov 27 04:49:25 localhost openstack_network_exporter[244641]: ERROR 09:49:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:49:25 localhost openstack_network_exporter[244641]: Nov 27 04:49:26 localhost nova_compute[284026]: 2025-11-27 09:49:26.333 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:27 localhost nova_compute[284026]: 2025-11-27 09:49:27.300 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:30 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48519 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A227DB0000000001030307) Nov 27 04:49:31 localhost nova_compute[284026]: 2025-11-27 09:49:31.336 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:32 localhost nova_compute[284026]: 2025-11-27 09:49:32.303 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:49:36 localhost podman[287738]: 2025-11-27 09:49:35.999735549 +0000 UTC m=+0.093134991 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:49:36 localhost podman[287738]: 2025-11-27 09:49:36.015109332 +0000 UTC m=+0.108508764 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, tcib_managed=true, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:49:36 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:49:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:49:36 localhost podman[287757]: 2025-11-27 09:49:36.134582852 +0000 UTC m=+0.075740605 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true) Nov 27 04:49:36 localhost podman[287757]: 2025-11-27 09:49:36.167836064 +0000 UTC m=+0.108993817 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:49:36 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:49:36 localhost nova_compute[284026]: 2025-11-27 09:49:36.338 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:37 localhost nova_compute[284026]: 2025-11-27 09:49:37.308 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:38 localhost podman[242678]: time="2025-11-27T09:49:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:49:38 localhost podman[242678]: @ - - [27/Nov/2025:09:49:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:49:38 localhost podman[242678]: @ - - [27/Nov/2025:09:49:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18214 "" "Go-http-client/1.1" Nov 27 04:49:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:49:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:49:39 localhost podman[287776]: 2025-11-27 09:49:39.997816933 +0000 UTC m=+0.096876189 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, config_id=edpm, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:49:40 localhost podman[287776]: 2025-11-27 09:49:40.009360577 +0000 UTC m=+0.108419853 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:49:40 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:49:40 localhost podman[287777]: 2025-11-27 09:49:40.105148166 +0000 UTC m=+0.201051539 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, distribution-scope=public, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=edpm_ansible, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, release=1755695350, version=9.6, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, io.buildah.version=1.33.7, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:49:40 localhost podman[287777]: 2025-11-27 09:49:40.141663863 +0000 UTC m=+0.237567246 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, container_name=openstack_network_exporter, managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7, release=1755695350, distribution-scope=public, com.redhat.component=ubi9-minimal-container, vcs-type=git, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=minimal rhel9, name=ubi9-minimal, io.openshift.expose-services=, config_id=edpm, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 04:49:40 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:49:41 localhost nova_compute[284026]: 2025-11-27 09:49:41.341 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:42 localhost nova_compute[284026]: 2025-11-27 09:49:42.340 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:49:43.558 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:49:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:49:43.558 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:49:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:49:43.559 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.134 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.135 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.139 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '983e6924-0070-4130-aa10-2496b778bfa6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.135571', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66e5ceba-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '23f20495f231c74d74571216278d8dd51b810574ab2ab34668d4f6ba7530630a'}]}, 'timestamp': '2025-11-27 09:49:44.140481', '_unique_id': 'b6bfae577c0b4c1d85610d2fddac1689'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.142 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.143 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.155 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.155 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '32d60ce6-672a-4ede-ae97-196c777cac7e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.143759', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66e81f80-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': 'e6e430daf1817becd26b91c612fb6edce38493001db938913e5025585be462b2'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.143759', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66e83222-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': 'bb19236581d41264221f1278889e0d0651b95bc4ad7dc578eb77a24ac948210c'}]}, 'timestamp': '2025-11-27 09:49:44.156008', '_unique_id': '4186f44ce36543eaabac1bc4d024d9d0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.157 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.158 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.190 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.190 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '44fdc177-4aba-4932-9dc0-7e2b5e721813', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.158283', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66ed83bc-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'fc1818e26eac6be945f2c668dba2fdf143a72251285a01476f766f194c7ab602'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.158283', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66ed98ca-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'bc3e7b7c874e2316f774bc25923f15f353080a13aa6c42dd71e4558531cd771a'}]}, 'timestamp': '2025-11-27 09:49:44.191422', '_unique_id': '090b9e63d3074104b76e16199fa0caee'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.194 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.194 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9fcdb1ca-9534-431c-9cd2-09b4c4ec937c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.194478', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66ee23bc-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '1972b4afe8cf5a620e2116f38536ae6188b3d3abb98e799cbf922b25c41e5913'}]}, 'timestamp': '2025-11-27 09:49:44.195067', '_unique_id': '541e9dd62d8e43a6878ff5e136433aa2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.197 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.197 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f352aa52-b016-47c3-b6c9-f453a68c6755', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.197222', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66ee8d16-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '29ef057aa001ee27e78387f5785ae431aba0910980380f05e14a03876bb085dd'}]}, 'timestamp': '2025-11-27 09:49:44.197721', '_unique_id': 'b88e718bbcd046cf8b374ece13f18c97'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.198 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.199 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.199 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.200 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd374f71e-d0c1-41d7-a789-7678f5ab61fe', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.199843', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66eef31e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '30fd4e4b95c66682456d89bbc736bc81717c456c36a43e8995ced2395e5f2fc5'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.199843', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66ef0322-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '8fde6fb84310ca848a4a6da61b55c957d777252cf2b2e388ba96b1dc086f349f'}]}, 'timestamp': '2025-11-27 09:49:44.200706', '_unique_id': '04b63687ad3a4ca3bc03d2ac32c26b3a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.201 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.202 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.202 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.203 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8c3c30e3-98b3-4129-87e8-d6cff3218cef', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.202883', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66ef69de-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '8e5d21eb66728c021d4ebd14ab81f305574ac7a3ec9d0b1afd19bb9c89e2104d'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.202883', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66ef7ae6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'b7c38105454fed4d72044ff1315632607cef9d1e7179c81409080a884354d9c2'}]}, 'timestamp': '2025-11-27 09:49:44.203747', '_unique_id': '2cfe90c2b8f34d74b16991c4bbb258a4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.205 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.205 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.206 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'acb51ec8-26f5-4e93-890d-28cb8c1e2081', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.205875', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66efdf90-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'e170282b19cae3f915f90273b9bea5525cea75162a5e0b30d818b050758a063f'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.205875', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66eff070-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '834506fb99e4f4891ed5195d496dba6bce5946d6f77b398e5a3b653432134302'}]}, 'timestamp': '2025-11-27 09:49:44.206753', '_unique_id': '69b64403258941c58e48839678245fc0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.208 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.208 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7c199881-200c-4d33-befe-215b37805859', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.208881', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f0543e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '0dbee790e4ff8642e9426842a04dfe763cf4e234f90776ea80af6fea6c3bd8b8'}]}, 'timestamp': '2025-11-27 09:49:44.209336', '_unique_id': 'ba67974113564409ac5ed88fdb6d08cb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.211 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c6e3a08e-7861-4c07-b991-9bc62942f607', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.211426', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f0b8de-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '1fe57df2d2343fca8c893155d84dedac50b48c9f5f95dbd96d94cdb131673702'}]}, 'timestamp': '2025-11-27 09:49:44.211911', '_unique_id': 'a4ef5577450442bab75623df0dd52d90'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.213 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '92621749-f917-47e8-9a3c-4566557b65ab', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.214004', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f11c48-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': 'fea9d87102f1b255fc23a6fadbb333093c26d28110a9678571dd5022d9a79bc9'}]}, 'timestamp': '2025-11-27 09:49:44.214486', '_unique_id': '16dc243f79634b628994be4a3e132b54'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.216 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.216 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd9bea94f-0852-42ae-abda-efb05afac78f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.216620', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66f1828c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '83f04caa950faf96b4defcb34c129ea695ede9d833bf18c94eadd505f34143de'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.216620', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66f19268-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': '0286b0a2ba5230d71d213996d36bd341c5826dd063a3febb4a052fe259125aaa'}]}, 'timestamp': '2025-11-27 09:49:44.217450', '_unique_id': '9096ec1d415146cab84c2351a0da9361'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.219 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.219 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '61d43620-a987-434c-9186-54ab995a45d4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.219704', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66f1fab4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'cf39d1a9c92f6eceb802c8bebda8a8c708beb928dc64110609a312a372ded3a9'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.219704', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66f20a90-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.346339618, 'message_signature': 'b828a2b97e5708a0b8981a35344f0ca0d41f7742bf0bfcda44c51108307a547a'}]}, 'timestamp': '2025-11-27 09:49:44.220559', '_unique_id': '8f2b6e85e01f4f3faa7249ec009e0e46'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.222 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.222 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b29a29ac-d580-49e1-964e-d775538e0dd9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.222695', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f26fbc-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '5f333c7a4c31e8d6f3f0673f2004ee9bb0485562f3ad3fb5bd43ed51dffa1b9b'}]}, 'timestamp': '2025-11-27 09:49:44.223144', '_unique_id': '747f89f561df459cabcaadea2be0d21e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.225 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 11160000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '82538bd6-7515-4246-a550-3c499f52ce4e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 11160000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:49:44.225208', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '66f50b96-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.427819392, 'message_signature': '50aeb357a905e0aa76dfe4e8e7bb5f70470b13b1d9b47240bc6858c9b66311c3'}]}, 'timestamp': '2025-11-27 09:49:44.240152', '_unique_id': 'bb4d7d9a06c241d1a12bc6dea8c131f8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.241 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eb9af785-aaa1-4d0b-9b5c-635a07d7abd1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.241452', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66f54a48-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': 'a38a511791f6b85eec3a830a7e87270d1f1d443b473695b0344b62af709c709b'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.241452', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66f55448-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': 'd66cb944572f785182f2fcc8d72be9a30cfbf2391290a53171ffefd15d3a069b'}]}, 'timestamp': '2025-11-27 09:49:44.242019', '_unique_id': '399be38aec484aed83136cdc33390c8b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.243 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.243 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.243 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.243 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4ddcf3a6-cf6c-46bc-aa1c-d883756b255d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:49:44.243424', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '66f596d8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': '3b7165665f4b5da77020f5c3781e4127c3e4ba7ef30723d6d8fcdbee3ec0aabd'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:49:44.243424', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '66f5a0f6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.331823267, 'message_signature': '9827a150285b244e005c40e7b963c18e26dfcc6148087db2fae36fba0f7b57be'}]}, 'timestamp': '2025-11-27 09:49:44.243962', '_unique_id': '9581c257ed99458292dbf849c7d661f5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'fbff1ff3-4e37-4617-a2ea-2b15cd6a80cf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:49:44.245331', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '66f5e098-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.427819392, 'message_signature': 'da71f3fc80dd603e9c139daf56f025fa91e035d754da31b49438eae94bdbad87'}]}, 'timestamp': '2025-11-27 09:49:44.245620', '_unique_id': 'eac966528ef24d88a3d90aaa1bd674d1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bb3404a7-ef56-4852-a96b-8c85e3bec5d2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.247131', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f62710-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': 'ad6c761a2300ab9d4745253d0dba590b3bdbac8cae0971519bf37bb5910dc680'}]}, 'timestamp': '2025-11-27 09:49:44.247415', '_unique_id': '8d88f339b6634849b6eb6df9240084ff'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'cf3831b7-afff-4954-91ad-9542d9726b3c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.248717', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f664fa-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': 'b4288f26ce9ad52ce83f3a86844890406edf977393e4dde000d573971066d28b'}]}, 'timestamp': '2025-11-27 09:49:44.249000', '_unique_id': 'c138f29f374944bf80bace042e4d32c8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.250 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e3076720-91ef-47df-a3b0-a36db954144d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:49:44.250369', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '66f6a564-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11214.323633972, 'message_signature': '80512310cc0c0276959cbcef24fcd5d89a7862b22fed7392d7e60dc725567010'}]}, 'timestamp': '2025-11-27 09:49:44.250918', '_unique_id': '9da31c7073d041769a0297d8d38b4c93'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:49:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:49:44.252 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:49:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:49:44 localhost podman[287814]: 2025-11-27 09:49:44.687489029 +0000 UTC m=+0.057936379 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:49:44 localhost podman[287814]: 2025-11-27 09:49:44.694614956 +0000 UTC m=+0.065062316 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:49:44 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:49:45 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20833 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A261A90000000001030307) Nov 27 04:49:46 localhost nova_compute[284026]: 2025-11-27 09:49:46.343 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:46 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20834 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2659A0000000001030307) Nov 27 04:49:47 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48520 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A267DA0000000001030307) Nov 27 04:49:47 localhost nova_compute[284026]: 2025-11-27 09:49:47.364 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:48 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20835 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A26D9A0000000001030307) Nov 27 04:49:49 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=25578 DF PROTO=TCP SPT=53004 DPT=9102 SEQ=1521050482 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A271DB0000000001030307) Nov 27 04:49:51 localhost nova_compute[284026]: 2025-11-27 09:49:51.345 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:52 localhost nova_compute[284026]: 2025-11-27 09:49:52.366 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:52 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20836 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A27D5B0000000001030307) Nov 27 04:49:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:49:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:49:52 localhost podman[287836]: 2025-11-27 09:49:52.995986591 +0000 UTC m=+0.087847672 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, org.label-schema.schema-version=1.0, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 04:49:53 localhost podman[287836]: 2025-11-27 09:49:53.031263756 +0000 UTC m=+0.123124777 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 04:49:53 localhost systemd[1]: tmp-crun.HMqqs4.mount: Deactivated successfully. Nov 27 04:49:53 localhost podman[287837]: 2025-11-27 09:49:53.050566862 +0000 UTC m=+0.137555375 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:49:53 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:49:53 localhost podman[287837]: 2025-11-27 09:49:53.087980522 +0000 UTC m=+0.174969065 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:49:53 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:49:55 localhost openstack_network_exporter[244641]: ERROR 09:49:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:49:55 localhost openstack_network_exporter[244641]: ERROR 09:49:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:49:55 localhost openstack_network_exporter[244641]: ERROR 09:49:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:49:55 localhost openstack_network_exporter[244641]: ERROR 09:49:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:49:55 localhost openstack_network_exporter[244641]: Nov 27 04:49:55 localhost openstack_network_exporter[244641]: ERROR 09:49:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:49:55 localhost openstack_network_exporter[244641]: Nov 27 04:49:56 localhost nova_compute[284026]: 2025-11-27 09:49:56.346 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:49:57 localhost nova_compute[284026]: 2025-11-27 09:49:57.402 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:00 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20837 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A29DDA0000000001030307) Nov 27 04:50:01 localhost nova_compute[284026]: 2025-11-27 09:50:01.349 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:02 localhost nova_compute[284026]: 2025-11-27 09:50:02.444 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:02 localhost sshd[287884]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:50:02 localhost systemd-logind[761]: New session 62 of user zuul. Nov 27 04:50:02 localhost systemd[1]: Started Session 62 of User zuul. Nov 27 04:50:02 localhost python3[287906]: ansible-ansible.legacy.command Invoked with _raw_params=subscription-manager unregister _uses_shell=True zuul_log_id=in-loop-ignore zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:50:03 localhost subscription-manager[287907]: Unregistered machine with identity: 88fcf67a-1c7b-4599-8a4d-ef1b8885a3ba Nov 27 04:50:03 localhost systemd-journald[49007]: Field hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.4 (251 of 333 items), suggesting rotation. Nov 27 04:50:03 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 04:50:03 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:50:03 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 04:50:06 localhost nova_compute[284026]: 2025-11-27 09:50:06.351 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:50:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:50:07 localhost podman[287911]: 2025-11-27 09:50:07.008973638 +0000 UTC m=+0.090211164 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, container_name=multipathd, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:50:07 localhost podman[287911]: 2025-11-27 09:50:07.021392228 +0000 UTC m=+0.102629734 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, container_name=multipathd, org.label-schema.license=GPLv2) Nov 27 04:50:07 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:50:07 localhost podman[287910]: 2025-11-27 09:50:07.11272537 +0000 UTC m=+0.193812582 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 04:50:07 localhost podman[287910]: 2025-11-27 09:50:07.121872833 +0000 UTC m=+0.202960045 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:50:07 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:50:07 localhost nova_compute[284026]: 2025-11-27 09:50:07.477 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:08 localhost podman[242678]: time="2025-11-27T09:50:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:50:08 localhost podman[242678]: @ - - [27/Nov/2025:09:50:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:50:08 localhost podman[242678]: @ - - [27/Nov/2025:09:50:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18222 "" "Go-http-client/1.1" Nov 27 04:50:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:50:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:50:10 localhost podman[287947]: 2025-11-27 09:50:10.962073373 +0000 UTC m=+0.064153093 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, config_id=edpm) Nov 27 04:50:10 localhost podman[287947]: 2025-11-27 09:50:10.973035943 +0000 UTC m=+0.075115683 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0) Nov 27 04:50:10 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:50:11 localhost systemd[1]: tmp-crun.tJ7DIi.mount: Deactivated successfully. Nov 27 04:50:11 localhost podman[287948]: 2025-11-27 09:50:11.017694647 +0000 UTC m=+0.116145211 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, url=https://catalog.redhat.com/en/search?searchType=containers, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, release=1755695350, maintainer=Red Hat, Inc., config_id=edpm, distribution-scope=public, architecture=x86_64, io.buildah.version=1.33.7, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=openstack_network_exporter, managed_by=edpm_ansible, io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, vcs-type=git, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal) Nov 27 04:50:11 localhost podman[287948]: 2025-11-27 09:50:11.031122264 +0000 UTC m=+0.129572858 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, distribution-scope=public, io.buildah.version=1.33.7, version=9.6, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, architecture=x86_64, config_id=edpm, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=ubi9-minimal-container, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, vcs-type=git, build-date=2025-08-20T13:12:41, release=1755695350) Nov 27 04:50:11 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:50:11 localhost nova_compute[284026]: 2025-11-27 09:50:11.353 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.332 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.333 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.333 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.333 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.501 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.585 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.586 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.586 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:50:12 localhost nova_compute[284026]: 2025-11-27 09:50:12.587 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.035 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.060 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.060 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.061 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.061 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.062 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.062 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.063 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.063 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.063 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.064 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.082 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.083 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.083 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.084 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.084 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.542 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.458s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.629 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.630 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.891 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.893 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11976MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.893 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:50:13 localhost nova_compute[284026]: 2025-11-27 09:50:13.894 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.018 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.019 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.020 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.065 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.521 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.456s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.528 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.575 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.578 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:50:14 localhost nova_compute[284026]: 2025-11-27 09:50:14.578 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.684s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:50:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:50:14 localhost podman[288117]: 2025-11-27 09:50:14.985794831 +0000 UTC m=+0.080642941 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:50:14 localhost podman[288117]: 2025-11-27 09:50:14.999007821 +0000 UTC m=+0.093855931 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:50:15 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:50:15 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43345 DF PROTO=TCP SPT=42762 DPT=9102 SEQ=2125142337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2D6D90000000001030307) Nov 27 04:50:16 localhost nova_compute[284026]: 2025-11-27 09:50:16.355 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:16 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43346 DF PROTO=TCP SPT=42762 DPT=9102 SEQ=2125142337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2DADA0000000001030307) Nov 27 04:50:17 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=20838 DF PROTO=TCP SPT=41762 DPT=9102 SEQ=3813011830 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2DDDB0000000001030307) Nov 27 04:50:17 localhost nova_compute[284026]: 2025-11-27 09:50:17.544 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:18 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43347 DF PROTO=TCP SPT=42762 DPT=9102 SEQ=2125142337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2E2DA0000000001030307) Nov 27 04:50:19 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=48521 DF PROTO=TCP SPT=41044 DPT=9102 SEQ=3674675545 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2E5DA0000000001030307) Nov 27 04:50:21 localhost nova_compute[284026]: 2025-11-27 09:50:21.357 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:22 localhost nova_compute[284026]: 2025-11-27 09:50:22.572 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:22 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43348 DF PROTO=TCP SPT=42762 DPT=9102 SEQ=2125142337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A2F29A0000000001030307) Nov 27 04:50:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:50:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:50:23 localhost podman[288142]: 2025-11-27 09:50:23.993010936 +0000 UTC m=+0.084956244 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:50:24 localhost podman[288142]: 2025-11-27 09:50:24.032876203 +0000 UTC m=+0.124821491 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:50:24 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:50:24 localhost podman[288143]: 2025-11-27 09:50:24.100620681 +0000 UTC m=+0.188664016 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:50:24 localhost podman[288143]: 2025-11-27 09:50:24.111842929 +0000 UTC m=+0.199886294 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:50:24 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:50:25 localhost openstack_network_exporter[244641]: ERROR 09:50:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:50:25 localhost openstack_network_exporter[244641]: ERROR 09:50:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:50:25 localhost openstack_network_exporter[244641]: ERROR 09:50:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:50:25 localhost openstack_network_exporter[244641]: ERROR 09:50:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:50:25 localhost openstack_network_exporter[244641]: Nov 27 04:50:25 localhost openstack_network_exporter[244641]: ERROR 09:50:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:50:25 localhost openstack_network_exporter[244641]: Nov 27 04:50:26 localhost nova_compute[284026]: 2025-11-27 09:50:26.359 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:26 localhost systemd[1]: virtsecretd.service: Deactivated successfully. Nov 27 04:50:27 localhost nova_compute[284026]: 2025-11-27 09:50:27.601 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:31 localhost kernel: DROPPING: IN=br-ex OUT= MACSRC=fa:16:3e:e9:55:a6 MACDST=fa:16:3e:cd:69:4f MACPROTO=0800 SRC=192.168.122.10 DST=192.168.122.108 LEN=60 TOS=0x00 PREC=0x00 TTL=62 ID=43349 DF PROTO=TCP SPT=42762 DPT=9102 SEQ=2125142337 ACK=0 WINDOW=32640 RES=0x00 SYN URGP=0 OPT (020405500402080A12A313DA0000000001030307) Nov 27 04:50:31 localhost nova_compute[284026]: 2025-11-27 09:50:31.360 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:32 localhost nova_compute[284026]: 2025-11-27 09:50:32.631 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:36 localhost nova_compute[284026]: 2025-11-27 09:50:36.363 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:37 localhost nova_compute[284026]: 2025-11-27 09:50:37.668 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:50:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:50:37 localhost podman[288244]: 2025-11-27 09:50:37.994636647 +0000 UTC m=+0.082990922 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:50:37 localhost podman[288244]: 2025-11-27 09:50:37.999094686 +0000 UTC m=+0.087448991 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:50:38 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:50:38 localhost podman[288245]: 2025-11-27 09:50:38.045558558 +0000 UTC m=+0.129574558 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:50:38 localhost podman[288245]: 2025-11-27 09:50:38.086947396 +0000 UTC m=+0.170963346 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:50:38 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:50:38 localhost podman[242678]: time="2025-11-27T09:50:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:50:38 localhost podman[242678]: @ - - [27/Nov/2025:09:50:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 151173 "" "Go-http-client/1.1" Nov 27 04:50:38 localhost podman[242678]: @ - - [27/Nov/2025:09:50:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18226 "" "Go-http-client/1.1" Nov 27 04:50:39 localhost sshd[288281]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:50:39 localhost systemd[1]: Created slice User Slice of UID 1003. Nov 27 04:50:39 localhost systemd[1]: Starting User Runtime Directory /run/user/1003... Nov 27 04:50:39 localhost systemd-logind[761]: New session 63 of user tripleo-admin. Nov 27 04:50:39 localhost systemd[1]: Finished User Runtime Directory /run/user/1003. Nov 27 04:50:39 localhost systemd[1]: Starting User Manager for UID 1003... Nov 27 04:50:39 localhost systemd[288285]: Queued start job for default target Main User Target. Nov 27 04:50:39 localhost systemd[288285]: Created slice User Application Slice. Nov 27 04:50:39 localhost systemd[288285]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 04:50:39 localhost systemd[288285]: Started Daily Cleanup of User's Temporary Directories. Nov 27 04:50:39 localhost systemd[288285]: Reached target Paths. Nov 27 04:50:39 localhost systemd[288285]: Reached target Timers. Nov 27 04:50:39 localhost systemd[288285]: Starting D-Bus User Message Bus Socket... Nov 27 04:50:39 localhost systemd[288285]: Starting Create User's Volatile Files and Directories... Nov 27 04:50:39 localhost systemd[288285]: Listening on D-Bus User Message Bus Socket. Nov 27 04:50:39 localhost systemd[288285]: Reached target Sockets. Nov 27 04:50:39 localhost systemd[288285]: Finished Create User's Volatile Files and Directories. Nov 27 04:50:39 localhost systemd[288285]: Reached target Basic System. Nov 27 04:50:39 localhost systemd[288285]: Reached target Main User Target. Nov 27 04:50:39 localhost systemd[288285]: Startup finished in 166ms. Nov 27 04:50:39 localhost systemd[1]: Started User Manager for UID 1003. Nov 27 04:50:39 localhost systemd[1]: Started Session 63 of User tripleo-admin. Nov 27 04:50:40 localhost python3[288428]: ansible-ansible.builtin.blockinfile Invoked with marker_begin=BEGIN ceph firewall rules marker_end=END ceph firewall rules path=/etc/nftables/edpm-rules.nft mode=0644 block=# 100 ceph_alertmanager (9093)#012add rule inet filter EDPM_INPUT tcp dport { 9093 } ct state new counter accept comment "100 ceph_alertmanager"#012# 100 ceph_dashboard (8443)#012add rule inet filter EDPM_INPUT tcp dport { 8443 } ct state new counter accept comment "100 ceph_dashboard"#012# 100 ceph_grafana (3100)#012add rule inet filter EDPM_INPUT tcp dport { 3100 } ct state new counter accept comment "100 ceph_grafana"#012# 100 ceph_prometheus (9092)#012add rule inet filter EDPM_INPUT tcp dport { 9092 } ct state new counter accept comment "100 ceph_prometheus"#012# 100 ceph_rgw (8080)#012add rule inet filter EDPM_INPUT tcp dport { 8080 } ct state new counter accept comment "100 ceph_rgw"#012# 110 ceph_mon (6789, 3300, 9100)#012add rule inet filter EDPM_INPUT tcp dport { 6789,3300,9100 } ct state new counter accept comment "110 ceph_mon"#012# 112 ceph_mds (6800-7300, 9100)#012add rule inet filter EDPM_INPUT tcp dport { 6800-7300,9100 } ct state new counter accept comment "112 ceph_mds"#012# 113 ceph_mgr (6800-7300, 8444)#012add rule inet filter EDPM_INPUT tcp dport { 6800-7300,8444 } ct state new counter accept comment "113 ceph_mgr"#012# 120 ceph_nfs (2049, 12049)#012add rule inet filter EDPM_INPUT tcp dport { 2049,12049 } ct state new counter accept comment "120 ceph_nfs"#012# 123 ceph_dashboard (9090, 9094, 9283)#012add rule inet filter EDPM_INPUT tcp dport { 9090,9094,9283 } ct state new counter accept comment "123 ceph_dashboard"#012 insertbefore=^# Lock down INPUT chains state=present marker=# {mark} ANSIBLE MANAGED BLOCK create=False backup=False unsafe_writes=False insertafter=None validate=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:50:41 localhost python3[288572]: ansible-ansible.builtin.systemd Invoked with name=nftables state=restarted enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Nov 27 04:50:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:50:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:50:41 localhost systemd[1]: Stopping Netfilter Tables... Nov 27 04:50:41 localhost podman[288574]: 2025-11-27 09:50:41.161102806 +0000 UTC m=+0.097686393 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, managed_by=edpm_ansible, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:50:41 localhost podman[288574]: 2025-11-27 09:50:41.169651172 +0000 UTC m=+0.106234779 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, io.buildah.version=1.41.3, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0) Nov 27 04:50:41 localhost systemd[1]: nftables.service: Deactivated successfully. Nov 27 04:50:41 localhost systemd[1]: Stopped Netfilter Tables. Nov 27 04:50:41 localhost systemd[1]: Starting Netfilter Tables... Nov 27 04:50:41 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:50:41 localhost systemd[1]: Finished Netfilter Tables. Nov 27 04:50:41 localhost podman[288575]: 2025-11-27 09:50:41.271699069 +0000 UTC m=+0.205763099 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, distribution-scope=public, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., version=9.6, name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=) Nov 27 04:50:41 localhost podman[288575]: 2025-11-27 09:50:41.290295883 +0000 UTC m=+0.224359873 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., maintainer=Red Hat, Inc., version=9.6, container_name=openstack_network_exporter, release=1755695350, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, managed_by=edpm_ansible, io.openshift.expose-services=, architecture=x86_64, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 04:50:41 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:50:41 localhost nova_compute[284026]: 2025-11-27 09:50:41.365 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:42 localhost nova_compute[284026]: 2025-11-27 09:50:42.703 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:50:43.559 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:50:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:50:43.560 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:50:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:50:43.561 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:50:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:50:45 localhost systemd[1]: tmp-crun.v8RfEm.mount: Deactivated successfully. Nov 27 04:50:45 localhost podman[288670]: 2025-11-27 09:50:45.385215308 +0000 UTC m=+0.088931169 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:50:45 localhost podman[288670]: 2025-11-27 09:50:45.398958053 +0000 UTC m=+0.102673954 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:50:45 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:50:46 localhost nova_compute[284026]: 2025-11-27 09:50:46.367 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:47 localhost nova_compute[284026]: 2025-11-27 09:50:47.706 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:51 localhost nova_compute[284026]: 2025-11-27 09:50:51.369 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:52 localhost nova_compute[284026]: 2025-11-27 09:50:52.740 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:52 localhost podman[288826]: Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.81314512 +0000 UTC m=+0.110725019 container create 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, vendor=Red Hat, Inc., io.buildah.version=1.33.12, GIT_CLEAN=True, GIT_BRANCH=main, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, io.openshift.expose-services=, ceph=True, io.openshift.tags=rhceph ceph, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, vcs-type=git, build-date=2025-09-24T08:57:55, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7) Nov 27 04:50:52 localhost systemd[1]: Started libpod-conmon-03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c.scope. Nov 27 04:50:52 localhost systemd[1]: Started libcrun container. Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.777867883 +0000 UTC m=+0.075447832 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.885840938 +0000 UTC m=+0.183420837 container init 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, RELEASE=main, name=rhceph, vcs-type=git, CEPH_POINT_RELEASE=, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.openshift.tags=rhceph ceph, architecture=x86_64, ceph=True, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, release=553, description=Red Hat Ceph Storage 7) Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.905565241 +0000 UTC m=+0.203145140 container start 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, GIT_BRANCH=main, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, version=7, io.buildah.version=1.33.12, vcs-type=git, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph) Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.906007893 +0000 UTC m=+0.203587832 container attach 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, RELEASE=main, vcs-type=git, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, GIT_BRANCH=main, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, version=7, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , io.openshift.expose-services=, architecture=x86_64, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:50:52 localhost sweet_galois[288841]: 167 167 Nov 27 04:50:52 localhost systemd[1]: libpod-03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c.scope: Deactivated successfully. Nov 27 04:50:52 localhost podman[288826]: 2025-11-27 09:50:52.910195764 +0000 UTC m=+0.207775693 container died 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, name=rhceph, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, version=7, GIT_BRANCH=main, io.openshift.expose-services=, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, RELEASE=main) Nov 27 04:50:53 localhost podman[288846]: 2025-11-27 09:50:53.009297283 +0000 UTC m=+0.090372238 container remove 03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_galois, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, GIT_CLEAN=True, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, io.openshift.expose-services=, com.redhat.component=rhceph-container, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, ceph=True, RELEASE=main, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, distribution-scope=public) Nov 27 04:50:53 localhost systemd[1]: libpod-conmon-03a5f9ac8c0c225579fea64c550f805b36df2a4c99ecbe182c23ecc42623019c.scope: Deactivated successfully. Nov 27 04:50:53 localhost systemd[1]: Reloading. Nov 27 04:50:53 localhost systemd-sysv-generator[288893]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:50:53 localhost systemd-rc-local-generator[288888]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: tmp-crun.V2JDxL.mount: Deactivated successfully. Nov 27 04:50:53 localhost systemd[1]: var-lib-containers-storage-overlay-3cc5ac44ebeabdbb0af4b3e2debbdf53f6bd61e3a727ddce176151dffbc4a188-merged.mount: Deactivated successfully. Nov 27 04:50:53 localhost systemd[1]: Reloading. Nov 27 04:50:53 localhost systemd-rc-local-generator[288925]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:50:53 localhost systemd-sysv-generator[288931]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:50:53 localhost systemd[1]: Starting Ceph mds.mds.np0005537446.hkzfou for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 04:50:54 localhost podman[288991]: Nov 27 04:50:54 localhost podman[288991]: 2025-11-27 09:50:54.175420826 +0000 UTC m=+0.078919564 container create e992309e1bf4b4c99a475b98c85d2df450cfb0d726a93732a829c887db7d39b1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mds-mds-np0005537446-hkzfou, ceph=True, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, release=553, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, name=rhceph, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, com.redhat.component=rhceph-container, version=7, RELEASE=main, GIT_CLEAN=True, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:50:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:50:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:50:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/221f4c5fe050849c1740a74e9c9ef4f5f9bf506801c3c33ced372d87c17acf7c/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:50:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/221f4c5fe050849c1740a74e9c9ef4f5f9bf506801c3c33ced372d87c17acf7c/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 04:50:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/221f4c5fe050849c1740a74e9c9ef4f5f9bf506801c3c33ced372d87c17acf7c/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:50:54 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/221f4c5fe050849c1740a74e9c9ef4f5f9bf506801c3c33ced372d87c17acf7c/merged/var/lib/ceph/mds/ceph-mds.np0005537446.hkzfou supports timestamps until 2038 (0x7fffffff) Nov 27 04:50:54 localhost podman[288991]: 2025-11-27 09:50:54.234472283 +0000 UTC m=+0.137971021 container init e992309e1bf4b4c99a475b98c85d2df450cfb0d726a93732a829c887db7d39b1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mds-mds-np0005537446-hkzfou, vendor=Red Hat, Inc., version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , RELEASE=main, com.redhat.component=rhceph-container, release=553, architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, GIT_CLEAN=True, ceph=True, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:50:54 localhost podman[288991]: 2025-11-27 09:50:54.14161764 +0000 UTC m=+0.045116428 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:50:54 localhost podman[288991]: 2025-11-27 09:50:54.245873015 +0000 UTC m=+0.149371753 container start e992309e1bf4b4c99a475b98c85d2df450cfb0d726a93732a829c887db7d39b1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mds-mds-np0005537446-hkzfou, architecture=x86_64, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., version=7, maintainer=Guillaume Abrioux , name=rhceph, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, ceph=True, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, distribution-scope=public, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main) Nov 27 04:50:54 localhost bash[288991]: e992309e1bf4b4c99a475b98c85d2df450cfb0d726a93732a829c887db7d39b1 Nov 27 04:50:54 localhost systemd[1]: Started Ceph mds.mds.np0005537446.hkzfou for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 04:50:54 localhost podman[289005]: 2025-11-27 09:50:54.300795463 +0000 UTC m=+0.083524787 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible) Nov 27 04:50:54 localhost ceph-mds[289027]: set uid:gid to 167:167 (ceph:ceph) Nov 27 04:50:54 localhost ceph-mds[289027]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-mds, pid 2 Nov 27 04:50:54 localhost ceph-mds[289027]: main not setting numa affinity Nov 27 04:50:54 localhost ceph-mds[289027]: pidfile_write: ignore empty --pid-file Nov 27 04:50:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mds-mds-np0005537446-hkzfou[289006]: starting mds.mds.np0005537446.hkzfou at Nov 27 04:50:54 localhost ceph-mds[289027]: mds.mds.np0005537446.hkzfou Updating MDS map to version 7 from mon.0 Nov 27 04:50:54 localhost podman[289005]: 2025-11-27 09:50:54.365004836 +0000 UTC m=+0.147734140 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 04:50:54 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:50:54 localhost podman[289008]: 2025-11-27 09:50:54.368819187 +0000 UTC m=+0.150687068 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:50:54 localhost podman[289008]: 2025-11-27 09:50:54.44958872 +0000 UTC m=+0.231456631 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:50:54 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:50:54 localhost systemd[1]: tmp-crun.JWAmBn.mount: Deactivated successfully. Nov 27 04:50:55 localhost ceph-mds[289027]: mds.mds.np0005537446.hkzfou Updating MDS map to version 8 from mon.0 Nov 27 04:50:55 localhost ceph-mds[289027]: mds.mds.np0005537446.hkzfou Monitors have assigned me to become a standby. Nov 27 04:50:55 localhost openstack_network_exporter[244641]: ERROR 09:50:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:50:55 localhost openstack_network_exporter[244641]: ERROR 09:50:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:50:55 localhost openstack_network_exporter[244641]: ERROR 09:50:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:50:55 localhost openstack_network_exporter[244641]: ERROR 09:50:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:50:55 localhost openstack_network_exporter[244641]: Nov 27 04:50:55 localhost openstack_network_exporter[244641]: ERROR 09:50:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:50:55 localhost openstack_network_exporter[244641]: Nov 27 04:50:56 localhost nova_compute[284026]: 2025-11-27 09:50:56.371 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:50:57 localhost nova_compute[284026]: 2025-11-27 09:50:57.777 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:00 localhost podman[289200]: 2025-11-27 09:51:00.349553969 +0000 UTC m=+0.092948986 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, com.redhat.component=rhceph-container, ceph=True, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, distribution-scope=public, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, version=7, vendor=Red Hat, Inc., name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=) Nov 27 04:51:00 localhost podman[289200]: 2025-11-27 09:51:00.450500406 +0000 UTC m=+0.193895433 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vendor=Red Hat, Inc., name=rhceph, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, release=553, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, ceph=True, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, distribution-scope=public, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:51:01 localhost nova_compute[284026]: 2025-11-27 09:51:01.372 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:02 localhost nova_compute[284026]: 2025-11-27 09:51:02.811 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:03 localhost systemd[1]: session-62.scope: Deactivated successfully. Nov 27 04:51:03 localhost systemd-logind[761]: Session 62 logged out. Waiting for processes to exit. Nov 27 04:51:03 localhost systemd-logind[761]: Removed session 62. Nov 27 04:51:06 localhost nova_compute[284026]: 2025-11-27 09:51:06.376 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:07 localhost nova_compute[284026]: 2025-11-27 09:51:07.846 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:08 localhost podman[242678]: time="2025-11-27T09:51:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:51:08 localhost podman[242678]: @ - - [27/Nov/2025:09:51:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 153251 "" "Go-http-client/1.1" Nov 27 04:51:08 localhost podman[242678]: @ - - [27/Nov/2025:09:51:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18707 "" "Go-http-client/1.1" Nov 27 04:51:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:51:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:51:09 localhost systemd[1]: tmp-crun.jXadyP.mount: Deactivated successfully. Nov 27 04:51:09 localhost podman[289324]: 2025-11-27 09:51:09.014757331 +0000 UTC m=+0.094157369 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 04:51:09 localhost podman[289324]: 2025-11-27 09:51:09.025615949 +0000 UTC m=+0.105015907 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:51:09 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:51:09 localhost podman[289323]: 2025-11-27 09:51:09.109655058 +0000 UTC m=+0.191787168 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 04:51:09 localhost podman[289323]: 2025-11-27 09:51:09.114487967 +0000 UTC m=+0.196620087 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, managed_by=edpm_ansible) Nov 27 04:51:09 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:51:11 localhost nova_compute[284026]: 2025-11-27 09:51:11.378 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:51:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:51:11 localhost systemd[1]: tmp-crun.QFTQgA.mount: Deactivated successfully. Nov 27 04:51:11 localhost podman[289361]: 2025-11-27 09:51:11.97590176 +0000 UTC m=+0.071332542 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, config_id=edpm, maintainer=Red Hat, Inc., managed_by=edpm_ansible, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.buildah.version=1.33.7, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=openstack_network_exporter, vcs-type=git, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, version=9.6, architecture=x86_64, distribution-scope=public) Nov 27 04:51:11 localhost podman[289361]: 2025-11-27 09:51:11.992329906 +0000 UTC m=+0.087760678 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, architecture=x86_64, distribution-scope=public, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., container_name=openstack_network_exporter, maintainer=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, name=ubi9-minimal, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal) Nov 27 04:51:12 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:51:12 localhost systemd[1]: tmp-crun.NBfst9.mount: Deactivated successfully. Nov 27 04:51:12 localhost podman[289360]: 2025-11-27 09:51:12.092881484 +0000 UTC m=+0.191211783 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, config_id=edpm, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:51:12 localhost podman[289360]: 2025-11-27 09:51:12.127901723 +0000 UTC m=+0.226232052 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS) Nov 27 04:51:12 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:51:12 localhost nova_compute[284026]: 2025-11-27 09:51:12.891 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:12 localhost nova_compute[284026]: 2025-11-27 09:51:12.973 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:12 localhost nova_compute[284026]: 2025-11-27 09:51:12.973 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.002 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.003 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.003 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.328 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.328 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.329 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.330 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.787 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.805 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.805 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.806 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.807 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.808 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.808 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.809 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.810 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.810 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.811 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.842 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.843 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.843 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.844 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:51:13 localhost nova_compute[284026]: 2025-11-27 09:51:13.845 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.356 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.511s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.424 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.424 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.651 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.653 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11965MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.654 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:51:14 localhost nova_compute[284026]: 2025-11-27 09:51:14.654 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.053 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.054 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.054 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.136 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.593 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.457s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.600 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.624 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.626 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:51:15 localhost nova_compute[284026]: 2025-11-27 09:51:15.626 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.972s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:51:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:51:15 localhost podman[289443]: 2025-11-27 09:51:15.984840507 +0000 UTC m=+0.077362443 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:51:15 localhost podman[289443]: 2025-11-27 09:51:15.997999445 +0000 UTC m=+0.090521391 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:51:16 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:51:16 localhost nova_compute[284026]: 2025-11-27 09:51:16.380 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:17 localhost nova_compute[284026]: 2025-11-27 09:51:17.914 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:21 localhost nova_compute[284026]: 2025-11-27 09:51:21.382 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:22 localhost nova_compute[284026]: 2025-11-27 09:51:22.919 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:51:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:51:24 localhost podman[289466]: 2025-11-27 09:51:24.997833607 +0000 UTC m=+0.087828601 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, managed_by=edpm_ansible, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, container_name=ovn_controller, org.label-schema.vendor=CentOS) Nov 27 04:51:25 localhost podman[289467]: 2025-11-27 09:51:25.052412374 +0000 UTC m=+0.140592710 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:51:25 localhost podman[289466]: 2025-11-27 09:51:25.070074283 +0000 UTC m=+0.160069257 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 04:51:25 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:51:25 localhost podman[289467]: 2025-11-27 09:51:25.091145962 +0000 UTC m=+0.179326258 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:51:25 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:51:25 localhost openstack_network_exporter[244641]: ERROR 09:51:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:51:25 localhost openstack_network_exporter[244641]: ERROR 09:51:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:51:25 localhost openstack_network_exporter[244641]: ERROR 09:51:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:51:25 localhost openstack_network_exporter[244641]: ERROR 09:51:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:51:25 localhost openstack_network_exporter[244641]: Nov 27 04:51:25 localhost openstack_network_exporter[244641]: ERROR 09:51:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:51:25 localhost openstack_network_exporter[244641]: Nov 27 04:51:26 localhost nova_compute[284026]: 2025-11-27 09:51:26.384 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:27 localhost nova_compute[284026]: 2025-11-27 09:51:27.945 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:31 localhost nova_compute[284026]: 2025-11-27 09:51:31.386 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:32 localhost nova_compute[284026]: 2025-11-27 09:51:32.982 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:36 localhost nova_compute[284026]: 2025-11-27 09:51:36.388 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:38 localhost nova_compute[284026]: 2025-11-27 09:51:38.008 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:38 localhost podman[242678]: time="2025-11-27T09:51:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:51:38 localhost podman[242678]: @ - - [27/Nov/2025:09:51:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 153251 "" "Go-http-client/1.1" Nov 27 04:51:38 localhost podman[242678]: @ - - [27/Nov/2025:09:51:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18715 "" "Go-http-client/1.1" Nov 27 04:51:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:51:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:51:40 localhost podman[289599]: 2025-11-27 09:51:40.006901384 +0000 UTC m=+0.086934627 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:51:40 localhost podman[289599]: 2025-11-27 09:51:40.036885288 +0000 UTC m=+0.116918511 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:51:40 localhost podman[289600]: 2025-11-27 09:51:40.061644455 +0000 UTC m=+0.139877201 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=multipathd) Nov 27 04:51:40 localhost podman[289600]: 2025-11-27 09:51:40.080712961 +0000 UTC m=+0.158945707 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:51:40 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:51:40 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:51:40 localhost systemd[1]: session-63.scope: Deactivated successfully. Nov 27 04:51:40 localhost systemd[1]: session-63.scope: Consumed 1.283s CPU time. Nov 27 04:51:40 localhost systemd-logind[761]: Session 63 logged out. Waiting for processes to exit. Nov 27 04:51:40 localhost systemd-logind[761]: Removed session 63. Nov 27 04:51:41 localhost nova_compute[284026]: 2025-11-27 09:51:41.389 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:51:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:51:42 localhost podman[289637]: 2025-11-27 09:51:42.994647088 +0000 UTC m=+0.084988825 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, io.buildah.version=1.41.3, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.vendor=CentOS) Nov 27 04:51:43 localhost podman[289637]: 2025-11-27 09:51:43.052164295 +0000 UTC m=+0.142506032 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:51:43 localhost nova_compute[284026]: 2025-11-27 09:51:43.054 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:43 localhost systemd[1]: tmp-crun.ibPto5.mount: Deactivated successfully. Nov 27 04:51:43 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:51:43 localhost podman[289638]: 2025-11-27 09:51:43.07157041 +0000 UTC m=+0.161394533 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, vcs-type=git, com.redhat.component=ubi9-minimal-container, io.openshift.expose-services=, config_id=edpm, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, distribution-scope=public, name=ubi9-minimal, version=9.6, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal) Nov 27 04:51:43 localhost podman[289638]: 2025-11-27 09:51:43.090152502 +0000 UTC m=+0.179976665 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, config_id=edpm, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, vendor=Red Hat, Inc., io.buildah.version=1.33.7, maintainer=Red Hat, Inc., vcs-type=git, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, version=9.6, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, name=ubi9-minimal, release=1755695350, architecture=x86_64, io.openshift.expose-services=) Nov 27 04:51:43 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:51:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:51:43.560 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:51:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:51:43.561 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:51:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:51:43.562 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.136 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.137 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.137 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.201 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.202 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bbc46bef-3507-47c6-bfbb-bf3ce224fc56', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.137708', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae75cf14-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': 'b8d7694e034d3d443a6c4e6d72a3f8db3350f5a573f0b585fb1b9f5a25a7cec3'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.137708', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae75ea4e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': 'f6844ac4b39bad5af41e8b6567b56155bd0b5e901dad39abde516da56ac6213e'}]}, 'timestamp': '2025-11-27 09:51:44.203042', '_unique_id': '9354a1604d804a6e8e164e1eda3f181b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.207 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eb24b81c-4689-4839-98d0-0ca40a002089', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.207357', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae7743da-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '97ff62a832bbd629296d1abc65f7e5027c772ae204e56459582993d254a66996'}]}, 'timestamp': '2025-11-27 09:51:44.211903', '_unique_id': 'fd1d3b7ff1bd4a79b40cbf4604b3a5a9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.212 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.213 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.214 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '70722471-49f9-424b-8143-7889c3dd8313', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.214239', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae77b20c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '4a019158098623343b36f875eff5724b6aec7bff143e76e6775bb6484d852dbb'}]}, 'timestamp': '2025-11-27 09:51:44.214766', '_unique_id': 'a1b154963d184788a95c8554dc3171c2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.216 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.216 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8be71340-6b0c-488a-85eb-42a418e25971', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.217039', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae781f12-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '9607d51b1cf6a68ee460ac4141bbe7a1d9ed32e3d53cd87f932e0b7553192ef1'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.217039', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae7830c4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': 'c79e976fd73bbbb56b63004898631401bdbe7075f42a18669909f43eefb2853c'}]}, 'timestamp': '2025-11-27 09:51:44.217926', '_unique_id': 'c579a7471f4644e3bd280ae539aa8e68'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.219 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.231 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.231 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd27197b8-5cf6-4128-a3d2-b605a1901dac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.220115', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae7a4e86-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': '005a040da748a37128ebc3e419bb5025c6b21e687abf08ca44bfe4eb97dd9871'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.220115', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae7a5f52-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': '1cecf9dfc8c5195c3d6124622750584ab7d1923cc2bc0983f16e63e146c266e9'}]}, 'timestamp': '2025-11-27 09:51:44.232222', '_unique_id': '55a65d8f117147e4b8ba8b9790d3ec2d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.234 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '35e7cbdc-6c06-4714-b363-1fb821ce9f5c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.234415', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae7ac802-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': '0b1e5729baac6f928aad4314bb1ed7c7fce68e43241097798864c4326e3d1502'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.234415', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae7ad838-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': 'aa401cfbd4aa55219705e29672f86ccf7036b677ad7d9e0c867bd90829e9e7de'}]}, 'timestamp': '2025-11-27 09:51:44.235314', '_unique_id': '4948b0d3321e42c2bb79c71532b37a8d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.236 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.237 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.237 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '24ee94f9-6014-43c3-b962-00a8ec281911', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.237469', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae7b3e7c-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': 'ae57d7803a04845c6a3996d5200991dd4ebe0242e1ac6ce3ac3c73ad80c83be9'}]}, 'timestamp': '2025-11-27 09:51:44.237964', '_unique_id': 'd25942c33f6d4182a16f6023958548f0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.239 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.240 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6eb30d0e-1339-4c19-8601-54419e8464be', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.240090', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae7ba33a-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '4716e1f238b4984e9494c39def56e8e663ef39262d3826bf12b01853af9905b4'}]}, 'timestamp': '2025-11-27 09:51:44.240573', '_unique_id': 'ea377b7228954af78c3e9f56222bbdc1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.242 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4b8beae5-bdef-4a4a-a803-512800a40c9e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.242666', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae7c0b5e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '482d55a40c625cb04bc62cd8211b8fb760fc7f8810701d7aafbe15c525e18525'}]}, 'timestamp': '2025-11-27 09:51:44.243221', '_unique_id': '93d24b23638a4ec29b34965f3e5a9b86'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.265 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 11790000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6aa4dd98-776f-42d1-80d1-e567e538e52e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 11790000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:51:44.245499', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'ae7f6fc4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.45296134, 'message_signature': '0a61f65ab218cde074d14dc88f74f484035716cfefaee78afcd36cd9f922e621'}]}, 'timestamp': '2025-11-27 09:51:44.265359', '_unique_id': 'b0296c53e26b47a395dcb5e53faf1bab'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.266 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd216d8e9-8571-4b07-9690-51205fdc62bb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.266830', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae7fb470-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '5912002d61bef35951dad056d01bbbaf789d8771cb7eebd9de7973e016ce81f4'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.266830', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae7fbe70-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '26e4c9934182d7dba7741f0f1e832ad8cb5740d98ec6ec858c1315a19e92a9c3'}]}, 'timestamp': '2025-11-27 09:51:44.267348', '_unique_id': '8fd48cd255f946ec970073566d531957'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.268 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.268 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c6e23a4f-dfc7-4b9d-94d4-2e514dde9bc8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.268711', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae7ffe26-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '05e77ab78434f8faac04bfb4b90ffe0dfbbeeee6930f63bccdddfd6d2edcbbf9'}]}, 'timestamp': '2025-11-27 09:51:44.268996', '_unique_id': 'bb28312ab65343e5897201c14edeb432'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.270 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.270 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'cf7909f9-ce3e-4ebe-b83d-0d7e288c29a8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.270270', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae803ad0-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '862e77117a0aaad1c10742e49295da2c7e3ed7a0b36ef3b714d51cdd63624a5e'}]}, 'timestamp': '2025-11-27 09:51:44.270571', '_unique_id': '535b2a1351a04cd4a033e2f29d80bd24'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.271 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bc7aa0c2-6b18-40be-9608-01922fcd066c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.271875', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae807996-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '04da24ac68db9150de14dc26760ce7d7f2d343db324dea9ca2da281589828eba'}]}, 'timestamp': '2025-11-27 09:51:44.272157', '_unique_id': '1e9b1ded66234d268db58a842726d8df'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.272 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.273 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.273 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '897d5872-16d8-4b75-8a24-cbf7bc50de37', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.273426', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae80b794-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': '63d5fb5d5dc5f3f0508102f2a208a885ba2b6e0b84be6687040ff2427f310ffc'}]}, 'timestamp': '2025-11-27 09:51:44.273745', '_unique_id': '37a3581317b8420ebe47bbd620b3b9ec'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.274 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'dd38f622-6bf9-4c2a-a650-8e7e66b18afc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:51:44.275057', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'ae80f5e2-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.395457014, 'message_signature': 'c0255ec4061bde6c748c687953bcf481fcddaa86704c3d17320caf9761110293'}]}, 'timestamp': '2025-11-27 09:51:44.275340', '_unique_id': 'bceb99fe6d864073b24e3d1e11f4e263'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.275 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.276 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.276 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.276 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a3052be1-d1c4-4f8b-9d5c-5797e80faa48', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.276637', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae8133b8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': 'b91dad65e40746a1430a60e40a2f707f740fcbb3004e8c678186a2f97c8580b0'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.276637', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae813db8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.408173372, 'message_signature': '2158f6a63015c794f85c168fe09777989deb451cb91e3424c66d6b063170c1cf'}]}, 'timestamp': '2025-11-27 09:51:44.277161', '_unique_id': '67da161e8be340d3b13e50abb95ba3d7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.278 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.278 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.278 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.278 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '26895eb6-9a93-402a-b098-de43578d5566', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.278579', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae817ff8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '8d1e1ddbaf610536cd43f696604ea169b41bffc9680da946683a41236702fb59'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.278579', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae818a02-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': 'a5cc15effecae9386ced8ea03f51863eb99093f34920f92e98563f710212dd48'}]}, 'timestamp': '2025-11-27 09:51:44.279112', '_unique_id': '8cebf0cd39af4af8ace4b4fc0ba69b02'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.280 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.280 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.280 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6fca38e9-97aa-4b72-8cbc-d2b330be0a2f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.280656', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae81d070-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '4407ea251311d2de78f9a7e764e2f4eac53b00f8cff7419841318944442d6c55'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.280656', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae81da66-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': 'ca14f4d97072729e79286e3f9b1efa5a6dcff3e611c8f31cf0adb886b60bcc98'}]}, 'timestamp': '2025-11-27 09:51:44.281170', '_unique_id': '4926b105366d44d7b0c49c44a229e6ea'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.281 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.282 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.282 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.282 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '708f223a-a10c-485d-ba02-2d9b76cb94ea', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:51:44.282476', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'ae821918-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '237fa3ef116c9160e664c9810e6bf7c256557f7ace3bacafa380c99cc17e0bf6'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:51:44.282476', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'ae822318-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.325780365, 'message_signature': '49fea1388d27aaf3e3393fe350086f3bdb4d2b309718569824ef0a1b76cdeeeb'}]}, 'timestamp': '2025-11-27 09:51:44.283030', '_unique_id': 'dad6bf9b73b042838c5595e4f84d60fd'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.283 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.284 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.284 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a37596ba-1706-4cbd-af0c-0c6862de9a90', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:51:44.284329', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'ae826008-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11334.45296134, 'message_signature': '4b7c7699917781c6c487c689fa7f08c555b8c4edcc375b6c8a771dc675dfb3f5'}]}, 'timestamp': '2025-11-27 09:51:44.284630', '_unique_id': '9aa7ceaa84d54db1b4e66a2d73b50139'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:51:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:51:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:51:46 localhost nova_compute[284026]: 2025-11-27 09:51:46.391 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:51:46 localhost podman[289676]: 2025-11-27 09:51:46.995256353 +0000 UTC m=+0.089255139 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:51:47 localhost podman[289676]: 2025-11-27 09:51:47.008496525 +0000 UTC m=+0.102495311 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:51:47 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:51:48 localhost nova_compute[284026]: 2025-11-27 09:51:48.095 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:50 localhost systemd[1]: Stopping User Manager for UID 1003... Nov 27 04:51:50 localhost systemd[288285]: Activating special unit Exit the Session... Nov 27 04:51:50 localhost systemd[288285]: Stopped target Main User Target. Nov 27 04:51:50 localhost systemd[288285]: Stopped target Basic System. Nov 27 04:51:50 localhost systemd[288285]: Stopped target Paths. Nov 27 04:51:50 localhost systemd[288285]: Stopped target Sockets. Nov 27 04:51:50 localhost systemd[288285]: Stopped target Timers. Nov 27 04:51:50 localhost systemd[288285]: Stopped Mark boot as successful after the user session has run 2 minutes. Nov 27 04:51:50 localhost systemd[288285]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 04:51:50 localhost systemd[288285]: Closed D-Bus User Message Bus Socket. Nov 27 04:51:50 localhost systemd[288285]: Stopped Create User's Volatile Files and Directories. Nov 27 04:51:50 localhost systemd[288285]: Removed slice User Application Slice. Nov 27 04:51:50 localhost systemd[288285]: Reached target Shutdown. Nov 27 04:51:50 localhost systemd[288285]: Finished Exit the Session. Nov 27 04:51:50 localhost systemd[288285]: Reached target Exit the Session. Nov 27 04:51:50 localhost systemd[1]: user@1003.service: Deactivated successfully. Nov 27 04:51:50 localhost systemd[1]: Stopped User Manager for UID 1003. Nov 27 04:51:50 localhost systemd[1]: Stopping User Runtime Directory /run/user/1003... Nov 27 04:51:50 localhost systemd[1]: run-user-1003.mount: Deactivated successfully. Nov 27 04:51:50 localhost systemd[1]: user-runtime-dir@1003.service: Deactivated successfully. Nov 27 04:51:50 localhost systemd[1]: Stopped User Runtime Directory /run/user/1003. Nov 27 04:51:50 localhost systemd[1]: Removed slice User Slice of UID 1003. Nov 27 04:51:50 localhost systemd[1]: user-1003.slice: Consumed 1.697s CPU time. Nov 27 04:51:51 localhost nova_compute[284026]: 2025-11-27 09:51:51.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:53 localhost nova_compute[284026]: 2025-11-27 09:51:53.142 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:55 localhost openstack_network_exporter[244641]: ERROR 09:51:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:51:55 localhost openstack_network_exporter[244641]: ERROR 09:51:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:51:55 localhost openstack_network_exporter[244641]: ERROR 09:51:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:51:55 localhost openstack_network_exporter[244641]: ERROR 09:51:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:51:55 localhost openstack_network_exporter[244641]: Nov 27 04:51:55 localhost openstack_network_exporter[244641]: ERROR 09:51:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:51:55 localhost openstack_network_exporter[244641]: Nov 27 04:51:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:51:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:51:56 localhost podman[289813]: 2025-11-27 09:51:55.999401438 +0000 UTC m=+0.088481868 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller) Nov 27 04:51:56 localhost podman[289814]: 2025-11-27 09:51:56.053098363 +0000 UTC m=+0.142715637 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:51:56 localhost podman[289814]: 2025-11-27 09:51:56.069904379 +0000 UTC m=+0.159521633 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:51:56 localhost podman[289813]: 2025-11-27 09:51:56.079735159 +0000 UTC m=+0.168815589 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, container_name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:51:56 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:51:56 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:51:56 localhost nova_compute[284026]: 2025-11-27 09:51:56.395 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:51:58 localhost nova_compute[284026]: 2025-11-27 09:51:58.187 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:01 localhost nova_compute[284026]: 2025-11-27 09:52:01.397 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:03 localhost nova_compute[284026]: 2025-11-27 09:52:03.235 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.751 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.753 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.754 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 04:52:05 localhost nova_compute[284026]: 2025-11-27 09:52:05.771 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:06 localhost nova_compute[284026]: 2025-11-27 09:52:06.399 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:07 localhost nova_compute[284026]: 2025-11-27 09:52:07.783 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:08 localhost nova_compute[284026]: 2025-11-27 09:52:08.261 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:08 localhost nova_compute[284026]: 2025-11-27 09:52:08.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:08 localhost podman[242678]: time="2025-11-27T09:52:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:52:08 localhost podman[242678]: @ - - [27/Nov/2025:09:52:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 153251 "" "Go-http-client/1.1" Nov 27 04:52:08 localhost podman[242678]: @ - - [27/Nov/2025:09:52:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 18715 "" "Go-http-client/1.1" Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.767 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.767 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.768 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.768 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:52:09 localhost nova_compute[284026]: 2025-11-27 09:52:09.769 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.244 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.475s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.304 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.305 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.516 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.517 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11958MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.517 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.518 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.633 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.634 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.634 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.707 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.723 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.724 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.737 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.775 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:52:10 localhost nova_compute[284026]: 2025-11-27 09:52:10.849 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:52:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:52:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:52:10 localhost podman[289886]: 2025-11-27 09:52:10.97859066 +0000 UTC m=+0.074560405 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent) Nov 27 04:52:11 localhost podman[289886]: 2025-11-27 09:52:11.015129243 +0000 UTC m=+0.111098998 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, io.buildah.version=1.41.3, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 04:52:11 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:52:11 localhost podman[289887]: 2025-11-27 09:52:11.071614631 +0000 UTC m=+0.155433480 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=multipathd, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:52:11 localhost podman[289887]: 2025-11-27 09:52:11.154336745 +0000 UTC m=+0.238155624 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd) Nov 27 04:52:11 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.343 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.495s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.351 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.367 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.369 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.369 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.851s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:52:11 localhost nova_compute[284026]: 2025-11-27 09:52:11.401 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:12 localhost nova_compute[284026]: 2025-11-27 09:52:12.369 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:12 localhost nova_compute[284026]: 2025-11-27 09:52:12.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:52:12 localhost nova_compute[284026]: 2025-11-27 09:52:12.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:52:12 localhost nova_compute[284026]: 2025-11-27 09:52:12.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:52:13 localhost nova_compute[284026]: 2025-11-27 09:52:13.264 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:13 localhost nova_compute[284026]: 2025-11-27 09:52:13.432 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:52:13 localhost nova_compute[284026]: 2025-11-27 09:52:13.433 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:52:13 localhost nova_compute[284026]: 2025-11-27 09:52:13.433 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:52:13 localhost nova_compute[284026]: 2025-11-27 09:52:13.433 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:52:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:52:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:52:13 localhost podman[289947]: 2025-11-27 09:52:13.987223463 +0000 UTC m=+0.070807794 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, maintainer=Red Hat, Inc., config_id=edpm, version=9.6, build-date=2025-08-20T13:12:41, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.33.7, container_name=openstack_network_exporter, vcs-type=git, name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:52:14 localhost podman[289947]: 2025-11-27 09:52:14.024894626 +0000 UTC m=+0.108478977 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, architecture=x86_64, io.openshift.tags=minimal rhel9, distribution-scope=public, maintainer=Red Hat, Inc., vcs-type=git, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, container_name=openstack_network_exporter, version=9.6, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, managed_by=edpm_ansible) Nov 27 04:52:14 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:52:14 localhost podman[289946]: 2025-11-27 09:52:14.040375442 +0000 UTC m=+0.129659366 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:52:14 localhost podman[289946]: 2025-11-27 09:52:14.050122105 +0000 UTC m=+0.139406049 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 04:52:14 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:52:14 localhost nova_compute[284026]: 2025-11-27 09:52:14.368 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:52:14 localhost nova_compute[284026]: 2025-11-27 09:52:14.383 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:52:14 localhost nova_compute[284026]: 2025-11-27 09:52:14.383 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:52:16 localhost nova_compute[284026]: 2025-11-27 09:52:16.404 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:52:18 localhost systemd[1]: tmp-crun.e88yh8.mount: Deactivated successfully. Nov 27 04:52:18 localhost podman[290004]: 2025-11-27 09:52:18.010798294 +0000 UTC m=+0.101422757 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:52:18 localhost podman[290004]: 2025-11-27 09:52:18.021934924 +0000 UTC m=+0.112559387 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:52:18 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:52:18 localhost nova_compute[284026]: 2025-11-27 09:52:18.297 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:21 localhost nova_compute[284026]: 2025-11-27 09:52:21.406 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:23 localhost nova_compute[284026]: 2025-11-27 09:52:23.299 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:25 localhost podman[290141]: Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.687668348 +0000 UTC m=+0.089578009 container create b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, name=rhceph, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, version=7, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, vcs-type=git, maintainer=Guillaume Abrioux , release=553, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:52:25 localhost systemd[1]: Started libpod-conmon-b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118.scope. Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.651853185 +0000 UTC m=+0.053762856 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:52:25 localhost systemd[1]: Started libcrun container. Nov 27 04:52:25 localhost openstack_network_exporter[244641]: ERROR 09:52:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:52:25 localhost openstack_network_exporter[244641]: ERROR 09:52:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:52:25 localhost openstack_network_exporter[244641]: ERROR 09:52:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:52:25 localhost openstack_network_exporter[244641]: ERROR 09:52:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:52:25 localhost openstack_network_exporter[244641]: Nov 27 04:52:25 localhost openstack_network_exporter[244641]: ERROR 09:52:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:52:25 localhost openstack_network_exporter[244641]: Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.789235258 +0000 UTC m=+0.191144919 container init b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, com.redhat.component=rhceph-container, version=7, RELEASE=main, GIT_BRANCH=main, release=553, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., vcs-type=git, name=rhceph, architecture=x86_64, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, GIT_CLEAN=True) Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.804059027 +0000 UTC m=+0.205968658 container start b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, GIT_CLEAN=True, architecture=x86_64, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, ceph=True, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, version=7, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=) Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.804331005 +0000 UTC m=+0.206240636 container attach b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, com.redhat.component=rhceph-container, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, version=7, RELEASE=main, architecture=x86_64, GIT_BRANCH=main, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7) Nov 27 04:52:25 localhost tender_shirley[290158]: 167 167 Nov 27 04:52:25 localhost systemd[1]: libpod-b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118.scope: Deactivated successfully. Nov 27 04:52:25 localhost podman[290141]: 2025-11-27 09:52:25.809111372 +0000 UTC m=+0.211021013 container died b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, GIT_BRANCH=main, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, ceph=True, distribution-scope=public, architecture=x86_64, io.buildah.version=1.33.12, release=553, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:52:25 localhost podman[290163]: 2025-11-27 09:52:25.917777634 +0000 UTC m=+0.096071544 container remove b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_shirley, GIT_BRANCH=main, CEPH_POINT_RELEASE=, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, name=rhceph, architecture=x86_64, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, io.openshift.expose-services=, GIT_CLEAN=True, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, vendor=Red Hat, Inc.) Nov 27 04:52:25 localhost systemd[1]: libpod-conmon-b6022dd5b0c9b989fbafe1d17808130fcd9cfae9b78e1eb391e7c58d37940118.scope: Deactivated successfully. Nov 27 04:52:25 localhost systemd[1]: Reloading. Nov 27 04:52:26 localhost systemd-rc-local-generator[290205]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:52:26 localhost systemd-sysv-generator[290210]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: var-lib-containers-storage-overlay-eaaf875fc37c5e5d3368b9bed6385487a6ccf4ea9a8303badc9c23b8f89e134b-merged.mount: Deactivated successfully. Nov 27 04:52:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:52:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:52:26 localhost nova_compute[284026]: 2025-11-27 09:52:26.408 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:26 localhost systemd[1]: Reloading. Nov 27 04:52:26 localhost podman[290220]: 2025-11-27 09:52:26.5049446 +0000 UTC m=+0.109562897 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:52:26 localhost podman[290220]: 2025-11-27 09:52:26.517821195 +0000 UTC m=+0.122439512 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:52:26 localhost systemd-sysv-generator[290285]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:52:26 localhost systemd-rc-local-generator[290281]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:52:26 localhost podman[290219]: 2025-11-27 09:52:26.598090514 +0000 UTC m=+0.203741809 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, container_name=ovn_controller, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller) Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:52:26 localhost podman[290219]: 2025-11-27 09:52:26.646593897 +0000 UTC m=+0.252245222 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.schema-version=1.0, container_name=ovn_controller, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:26 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:52:26 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:52:26 localhost systemd[1]: Starting Ceph mgr.np0005537446.nfulyg for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 04:52:27 localhost podman[290359]: Nov 27 04:52:27 localhost podman[290359]: 2025-11-27 09:52:27.182082704 +0000 UTC m=+0.077946976 container create 0ae9498f72e41a40351c1bf68ee106c7dea22dec32cd51fb6bf33b913ac26c93 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg, description=Red Hat Ceph Storage 7, vcs-type=git, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, release=553, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, version=7, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, ceph=True, GIT_BRANCH=main, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:52:27 localhost podman[290359]: 2025-11-27 09:52:27.145286194 +0000 UTC m=+0.041150526 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:52:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f7ffd279f35db596af0834b9fbeefcca5a3d1e0d3735fe8430f133507fd57d3b/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f7ffd279f35db596af0834b9fbeefcca5a3d1e0d3735fe8430f133507fd57d3b/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f7ffd279f35db596af0834b9fbeefcca5a3d1e0d3735fe8430f133507fd57d3b/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:27 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f7ffd279f35db596af0834b9fbeefcca5a3d1e0d3735fe8430f133507fd57d3b/merged/var/lib/ceph/mgr/ceph-np0005537446.nfulyg supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:27 localhost podman[290359]: 2025-11-27 09:52:27.260299316 +0000 UTC m=+0.156163588 container init 0ae9498f72e41a40351c1bf68ee106c7dea22dec32cd51fb6bf33b913ac26c93 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg, build-date=2025-09-24T08:57:55, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , GIT_CLEAN=True, RELEASE=main, ceph=True, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public) Nov 27 04:52:27 localhost podman[290359]: 2025-11-27 09:52:27.273385738 +0000 UTC m=+0.169250000 container start 0ae9498f72e41a40351c1bf68ee106c7dea22dec32cd51fb6bf33b913ac26c93 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg, io.openshift.expose-services=, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, release=553, maintainer=Guillaume Abrioux , architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, GIT_BRANCH=main, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, io.buildah.version=1.33.12, ceph=True, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:52:27 localhost bash[290359]: 0ae9498f72e41a40351c1bf68ee106c7dea22dec32cd51fb6bf33b913ac26c93 Nov 27 04:52:27 localhost systemd[1]: Started Ceph mgr.np0005537446.nfulyg for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 04:52:27 localhost ceph-mgr[290377]: set uid:gid to 167:167 (ceph:ceph) Nov 27 04:52:27 localhost ceph-mgr[290377]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-mgr, pid 2 Nov 27 04:52:27 localhost ceph-mgr[290377]: pidfile_write: ignore empty --pid-file Nov 27 04:52:27 localhost ceph-mgr[290377]: mgr[py] Loading python module 'alerts' Nov 27 04:52:27 localhost ceph-mgr[290377]: mgr[py] Module alerts has missing NOTIFY_TYPES member Nov 27 04:52:27 localhost ceph-mgr[290377]: mgr[py] Loading python module 'balancer' Nov 27 04:52:27 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:27.457+0000 7fae7e6da140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member Nov 27 04:52:27 localhost ceph-mgr[290377]: mgr[py] Module balancer has missing NOTIFY_TYPES member Nov 27 04:52:27 localhost ceph-mgr[290377]: mgr[py] Loading python module 'cephadm' Nov 27 04:52:27 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:27.527+0000 7fae7e6da140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member Nov 27 04:52:27 localhost systemd[1]: tmp-crun.D2vOR9.mount: Deactivated successfully. Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'crash' Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Module crash has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'dashboard' Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:28.245+0000 7fae7e6da140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost nova_compute[284026]: 2025-11-27 09:52:28.304 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'devicehealth' Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Module devicehealth has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'diskprediction_local' Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:28.789+0000 7fae7e6da140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: from numpy import show_config as show_numpy_config Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:28.921+0000 7fae7e6da140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'influx' Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Module influx has missing NOTIFY_TYPES member Nov 27 04:52:28 localhost ceph-mgr[290377]: mgr[py] Loading python module 'insights' Nov 27 04:52:28 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:28.981+0000 7fae7e6da140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost podman[290534]: 2025-11-27 09:52:29.029557142 +0000 UTC m=+0.109780033 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, version=7, GIT_BRANCH=main, architecture=x86_64, distribution-scope=public, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, vendor=Red Hat, Inc., GIT_CLEAN=True) Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'iostat' Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Module iostat has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'k8sevents' Nov 27 04:52:29 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:29.093+0000 7fae7e6da140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost podman[290534]: 2025-11-27 09:52:29.134010049 +0000 UTC m=+0.214232980 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, maintainer=Guillaume Abrioux , vcs-type=git, io.openshift.tags=rhceph ceph, name=rhceph, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, ceph=True, GIT_BRANCH=main, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container) Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'localpool' Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'mds_autoscaler' Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'mirroring' Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'nfs' Nov 27 04:52:29 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:29.832+0000 7fae7e6da140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Module nfs has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'orchestrator' Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Module orchestrator has missing NOTIFY_TYPES member Nov 27 04:52:29 localhost ceph-mgr[290377]: mgr[py] Loading python module 'osd_perf_query' Nov 27 04:52:29 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:29.973+0000 7fae7e6da140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'osd_support' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.036+0000 7fae7e6da140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module osd_support has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'pg_autoscaler' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.092+0000 7fae7e6da140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'progress' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.157+0000 7fae7e6da140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module progress has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'prometheus' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.216+0000 7fae7e6da140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module prometheus has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rbd_support' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.510+0000 7fae7e6da140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module rbd_support has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.590+0000 7fae7e6da140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'restful' Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rgw' Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Module rgw has missing NOTIFY_TYPES member Nov 27 04:52:30 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rook' Nov 27 04:52:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:30.914+0000 7fae7e6da140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module rook has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'selftest' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.331+0000 7fae7e6da140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module selftest has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'snap_schedule' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.391+0000 7fae7e6da140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost nova_compute[284026]: 2025-11-27 09:52:31.413 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'stats' Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'status' Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module status has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'telegraf' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.580+0000 7fae7e6da140 -1 mgr[py] Module status has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module telegraf has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'telemetry' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.637+0000 7fae7e6da140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module telemetry has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'test_orchestrator' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.768+0000 7fae7e6da140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member Nov 27 04:52:31 localhost ceph-mgr[290377]: mgr[py] Loading python module 'volumes' Nov 27 04:52:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:31.915+0000 7fae7e6da140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member Nov 27 04:52:32 localhost ceph-mgr[290377]: mgr[py] Module volumes has missing NOTIFY_TYPES member Nov 27 04:52:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:32.098+0000 7fae7e6da140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member Nov 27 04:52:32 localhost ceph-mgr[290377]: mgr[py] Loading python module 'zabbix' Nov 27 04:52:32 localhost ceph-mgr[290377]: mgr[py] Module zabbix has missing NOTIFY_TYPES member Nov 27 04:52:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:52:32.156+0000 7fae7e6da140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member Nov 27 04:52:32 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff551e0 mon_map magic: 0 from mon.1 v2:172.18.0.105:3300/0 Nov 27 04:52:32 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.103:6800/2529754064 Nov 27 04:52:33 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.103:6800/2529754064 Nov 27 04:52:33 localhost nova_compute[284026]: 2025-11-27 09:52:33.304 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:36 localhost nova_compute[284026]: 2025-11-27 09:52:36.415 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:38 localhost nova_compute[284026]: 2025-11-27 09:52:38.327 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:38 localhost podman[242678]: time="2025-11-27T09:52:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:52:38 localhost podman[242678]: @ - - [27/Nov/2025:09:52:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 155317 "" "Go-http-client/1.1" Nov 27 04:52:38 localhost podman[242678]: @ - - [27/Nov/2025:09:52:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19198 "" "Go-http-client/1.1" Nov 27 04:52:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:52:41 localhost podman[291414]: 2025-11-27 09:52:41.225206249 +0000 UTC m=+0.097810760 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:52:41 localhost podman[291414]: 2025-11-27 09:52:41.235199718 +0000 UTC m=+0.107804259 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:52:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:52:41 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:52:41 localhost podman[291449]: 2025-11-27 09:52:41.338773242 +0000 UTC m=+0.089006404 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:52:41 localhost podman[291449]: 2025-11-27 09:52:41.381201743 +0000 UTC m=+0.131434885 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 04:52:41 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:52:41 localhost nova_compute[284026]: 2025-11-27 09:52:41.417 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:41 localhost podman[291531]: Nov 27 04:52:41 localhost podman[291531]: 2025-11-27 09:52:41.984568714 +0000 UTC m=+0.065876802 container create 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, CEPH_POINT_RELEASE=, RELEASE=main, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , version=7, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_CLEAN=True, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:52:42 localhost systemd[1]: Started libpod-conmon-51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763.scope. Nov 27 04:52:42 localhost systemd[1]: Started libcrun container. Nov 27 04:52:42 localhost podman[291531]: 2025-11-27 09:52:41.954420064 +0000 UTC m=+0.035728182 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:52:42 localhost podman[291531]: 2025-11-27 09:52:42.102944126 +0000 UTC m=+0.184252214 container init 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, distribution-scope=public, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, name=rhceph, io.openshift.tags=rhceph ceph, GIT_BRANCH=main) Nov 27 04:52:42 localhost podman[291531]: 2025-11-27 09:52:42.112334179 +0000 UTC m=+0.193642267 container start 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, release=553, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, RELEASE=main, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., name=rhceph, GIT_CLEAN=True, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:52:42 localhost podman[291531]: 2025-11-27 09:52:42.112686398 +0000 UTC m=+0.193994536 container attach 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, version=7, distribution-scope=public, architecture=x86_64, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, release=553, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, vcs-type=git) Nov 27 04:52:42 localhost quizzical_johnson[291545]: 167 167 Nov 27 04:52:42 localhost systemd[1]: libpod-51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763.scope: Deactivated successfully. Nov 27 04:52:42 localhost podman[291531]: 2025-11-27 09:52:42.118029032 +0000 UTC m=+0.199337120 container died 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, release=553, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., GIT_CLEAN=True, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , distribution-scope=public, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, RELEASE=main, architecture=x86_64, GIT_BRANCH=main, CEPH_POINT_RELEASE=, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:52:42 localhost podman[291550]: 2025-11-27 09:52:42.215985375 +0000 UTC m=+0.086867536 container remove 51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_johnson, io.openshift.tags=rhceph ceph, release=553, name=rhceph, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, io.openshift.expose-services=, distribution-scope=public, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, ceph=True, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux ) Nov 27 04:52:42 localhost systemd[1]: var-lib-containers-storage-overlay-49fe62bf584d5c4de8203015840975173b7fe32c85524877cbfdeff00e70bdbe-merged.mount: Deactivated successfully. Nov 27 04:52:42 localhost systemd[1]: libpod-conmon-51db2b8247cff33ce6d63d409fa284bd9005b06d39e6e561c5e2ebaf305d0763.scope: Deactivated successfully. Nov 27 04:52:42 localhost podman[291566]: Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.343044802 +0000 UTC m=+0.081274516 container create 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, maintainer=Guillaume Abrioux , GIT_CLEAN=True, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, RELEASE=main, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, vendor=Red Hat, Inc., version=7, ceph=True, release=553, com.redhat.component=rhceph-container, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:52:42 localhost systemd[1]: Started libpod-conmon-28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c.scope. Nov 27 04:52:42 localhost systemd[1]: Started libcrun container. Nov 27 04:52:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6767a5c93555c450a19a56fe6a54319979380413cced18b66c453f3848e4aa0e/merged/tmp/config supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6767a5c93555c450a19a56fe6a54319979380413cced18b66c453f3848e4aa0e/merged/tmp/keyring supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6767a5c93555c450a19a56fe6a54319979380413cced18b66c453f3848e4aa0e/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6767a5c93555c450a19a56fe6a54319979380413cced18b66c453f3848e4aa0e/merged/var/lib/ceph/mon/ceph-np0005537446 supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.310698112 +0000 UTC m=+0.048927866 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.412085897 +0000 UTC m=+0.150315611 container init 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, RELEASE=main, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , release=553, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., version=7, GIT_BRANCH=main, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, name=rhceph, description=Red Hat Ceph Storage 7) Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.420196305 +0000 UTC m=+0.158426009 container start 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, release=553, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, name=rhceph, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, io.openshift.expose-services=, GIT_CLEAN=True, architecture=x86_64, ceph=True, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.420445602 +0000 UTC m=+0.158675366 container attach 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , RELEASE=main, GIT_BRANCH=main, ceph=True, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, architecture=x86_64, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553) Nov 27 04:52:42 localhost systemd[1]: libpod-28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c.scope: Deactivated successfully. Nov 27 04:52:42 localhost podman[291566]: 2025-11-27 09:52:42.522168286 +0000 UTC m=+0.260398020 container died 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, io.openshift.tags=rhceph ceph, name=rhceph, maintainer=Guillaume Abrioux , RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_BRANCH=main, version=7, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:52:42 localhost podman[291607]: 2025-11-27 09:52:42.617334725 +0000 UTC m=+0.083442264 container remove 28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=optimistic_poincare, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, distribution-scope=public, io.openshift.expose-services=, CEPH_POINT_RELEASE=, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, version=7, ceph=True) Nov 27 04:52:42 localhost systemd[1]: libpod-conmon-28214e7c762016c33669766069a1a213522b23eb118bd0590e8cc6a59cbda60c.scope: Deactivated successfully. Nov 27 04:52:42 localhost systemd[1]: Reloading. Nov 27 04:52:42 localhost systemd-sysv-generator[291648]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:52:42 localhost systemd-rc-local-generator[291643]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:42 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: var-lib-containers-storage-overlay-6767a5c93555c450a19a56fe6a54319979380413cced18b66c453f3848e4aa0e-merged.mount: Deactivated successfully. Nov 27 04:52:43 localhost systemd[1]: Reloading. Nov 27 04:52:43 localhost systemd-rc-local-generator[291691]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:52:43 localhost systemd-sysv-generator[291695]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:52:43 localhost nova_compute[284026]: 2025-11-27 09:52:43.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:52:43 localhost systemd[1]: Starting Ceph mon.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 04:52:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:52:43.562 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:52:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:52:43.563 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:52:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:52:43.564 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:52:43 localhost podman[291754]: Nov 27 04:52:43 localhost podman[291754]: 2025-11-27 09:52:43.880115884 +0000 UTC m=+0.077114214 container create f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, distribution-scope=public, maintainer=Guillaume Abrioux , GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:52:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d2767724e77723f2921128140e380c6e7b2790b261ea5617ea23aa84e7c9c784/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d2767724e77723f2921128140e380c6e7b2790b261ea5617ea23aa84e7c9c784/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d2767724e77723f2921128140e380c6e7b2790b261ea5617ea23aa84e7c9c784/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d2767724e77723f2921128140e380c6e7b2790b261ea5617ea23aa84e7c9c784/merged/var/lib/ceph/mon/ceph-np0005537446 supports timestamps until 2038 (0x7fffffff) Nov 27 04:52:43 localhost podman[291754]: 2025-11-27 09:52:43.943786776 +0000 UTC m=+0.140785056 container init f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, architecture=x86_64, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , name=rhceph, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, GIT_CLEAN=True, CEPH_POINT_RELEASE=, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, io.openshift.tags=rhceph ceph, version=7, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, vendor=Red Hat, Inc., io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main) Nov 27 04:52:43 localhost podman[291754]: 2025-11-27 09:52:43.848033842 +0000 UTC m=+0.045032192 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:52:43 localhost systemd[1]: tmp-crun.vkd8MX.mount: Deactivated successfully. Nov 27 04:52:43 localhost podman[291754]: 2025-11-27 09:52:43.953303191 +0000 UTC m=+0.150301491 container start f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, io.buildah.version=1.33.12, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, RELEASE=main, vendor=Red Hat, Inc., distribution-scope=public, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, vcs-type=git, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, version=7, architecture=x86_64, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:52:43 localhost bash[291754]: f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c Nov 27 04:52:43 localhost systemd[1]: Started Ceph mon.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 04:52:43 localhost ceph-mon[291772]: set uid:gid to 167:167 (ceph:ceph) Nov 27 04:52:43 localhost ceph-mon[291772]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-mon, pid 2 Nov 27 04:52:43 localhost ceph-mon[291772]: pidfile_write: ignore empty --pid-file Nov 27 04:52:43 localhost ceph-mon[291772]: load: jerasure load: lrc Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: RocksDB version: 7.9.2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Git sha 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: DB SUMMARY Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: DB Session ID: GVMC457KA1L6MF3IGLXZ Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: CURRENT file: CURRENT Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: IDENTITY file: IDENTITY Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: SST files in /var/lib/ceph/mon/ceph-np0005537446/store.db dir, Total Num: 0, files: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-np0005537446/store.db: 000004.log size: 761 ; Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.error_if_exists: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.create_if_missing: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.paranoid_checks: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.env: 0x5606459f69e0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.fs: PosixFileSystem Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.info_log: 0x56064731ed20 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.statistics: (nil) Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.use_fsync: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_log_file_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_fallocate: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.use_direct_reads: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.create_missing_column_families: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.db_log_dir: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.wal_dir: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.advise_random_on_open: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.write_buffer_manager: 0x56064732f540 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.rate_limiter: (nil) Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.unordered_write: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.row_cache: None Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.wal_filter: None Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.two_write_queues: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.manual_wal_flush: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.wal_compression: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.atomic_flush: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.log_readahead_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.db_host_id: __hostname__ Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_background_jobs: 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_background_compactions: -1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_subcompactions: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.writable_file_max_buffer_size: 1048576 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_total_wal_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_open_files: -1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bytes_per_sync: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_readahead_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_background_flushes: -1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Compression algorithms supported: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kZSTD supported: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kXpressCompression supported: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kZlibCompression supported: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-np0005537446/store.db/MANIFEST-000005 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.merge_operator: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_filter: None Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_filter_factory: None Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.sst_partitioner_factory: None Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x56064731e980)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x56064731b350#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.write_buffer_size: 33554432 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_write_buffer_number: 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression: NoCompression Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression: Disabled Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.prefix_extractor: nullptr Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.num_levels: 7 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.min_write_buffer_number_to_merge: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.level: 32767 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.enabled: false Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.level0_file_num_compaction_trigger: 4 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_base: 268435456 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.arena_block_size: 1048576 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.table_properties_collectors: Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.inplace_update_support: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.bloom_locality: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.max_successive_merges: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.force_consistency_checks: 1 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.ttl: 2592000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enable_blob_files: false Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.min_blob_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_file_size: 268435456 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-np0005537446/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: e15a37f6-4d8e-4c18-a493-be6024e004ec Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237163981334, "job": 1, "event": "recovery_started", "wal_files": [4]} Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237163983093, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1887, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 773, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 651, "raw_average_value_size": 130, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237163983179, "job": 1, "event": "recovery_finished"} Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x560647342e00 Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: DB pointer 0x560647438000 Nov 27 04:52:43 localhost ceph-mon[291772]: mon.np0005537446 does not exist in monmap, will attempt to join an existing cluster Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:52:43 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 0.0 total, 0.0 interval#012Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s#012Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 1/0 1.84 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 1.0 0.00 0.00 1 0.002 0 0 0.0 0.0#012 Sum 1/0 1.84 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 1.0 0.00 0.00 1 0.002 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 1.0 0.00 0.00 1 0.002 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.00 0.00 1 0.002 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.0 total, 0.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.17 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.17 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x56064731b350#2 capacity: 512.00 MB usage: 0.22 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 3.2e-05 secs_since: 0#012Block cache entry stats(count,size,portion): FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(2,0.95 KB,0.000181794%)#012#012** File Read Latency Histogram By Level [default] ** Nov 27 04:52:43 localhost ceph-mon[291772]: using public_addr v2:172.18.0.108:0/0 -> [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] Nov 27 04:52:43 localhost ceph-mon[291772]: starting mon.np0005537446 rank -1 at public addrs [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] at bind addrs [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] mon_data /var/lib/ceph/mon/ceph-np0005537446 fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:52:43 localhost ceph-mon[291772]: mon.np0005537446@-1(???) e0 preinit fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing) e3 sync_obtain_latest_monmap Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing) e3 sync_obtain_latest_monmap obtained monmap e3 Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).mds e17 new map Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).mds e17 print_map#012e17#012enable_multiple, ever_enabled_multiple: 1,1#012default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,12=quiesce subvolumes}#012legacy client fscid: 1#012 #012Filesystem 'cephfs' (1)#012fs_name#011cephfs#012epoch#01116#012flags#01112 joinable allow_snaps allow_multimds_snaps#012created#0112025-11-27T08:07:00.938837+0000#012modified#0112025-11-27T09:51:38.628845+0000#012tableserver#0110#012root#0110#012session_timeout#01160#012session_autoclose#011300#012max_file_size#0111099511627776#012required_client_features#011{}#012last_failure#0110#012last_failure_osd_epoch#01181#012compat#011compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2,12=quiesce subvolumes}#012max_mds#0111#012in#0110#012up#011{0=26518}#012failed#011#012damaged#011#012stopped#011#012data_pools#011[6]#012metadata_pool#0117#012inline_data#011disabled#012balancer#011#012bal_rank_mask#011-1#012standby_count_wanted#0111#012qdb_cluster#011leader: 26518 members: 26518#012[mds.mds.np0005537445.rkchqo{0:26518} state up:active seq 13 addr [v2:172.18.0.107:6808/3417254461,v1:172.18.0.107:6809/3417254461] compat {c=[1],r=[1],i=[17ff]}]#012 #012 #012Standby daemons:#012 #012[mds.mds.np0005537446.hkzfou{-1:17040} state up:standby seq 1 addr [v2:172.18.0.108:6808/2569123203,v1:172.18.0.108:6809/2569123203] compat {c=[1],r=[1],i=[17ff]}]#012[mds.mds.np0005537444.iqtbiy{-1:17052} state up:standby seq 1 addr [v2:172.18.0.106:6808/4180494082,v1:172.18.0.106:6809/4180494082] compat {c=[1],r=[1],i=[17ff]}] Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).osd e82 crush map has features 3314933000854323200, adjusting msgr requires Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).osd e82 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).osd e82 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).osd e82 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mgr to host np0005537444.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mgr to host np0005537445.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mgr to host np0005537446.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished Nov 27 04:52:44 localhost ceph-mon[291772]: Saving service mgr spec with placement label:mgr Nov 27 04:52:44 localhost ceph-mon[291772]: Deploying daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished Nov 27 04:52:44 localhost ceph-mon[291772]: Deploying daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537441.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537441.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished Nov 27 04:52:44 localhost ceph-mon[291772]: Deploying daemon mgr.np0005537446.nfulyg on np0005537446.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537442.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537442.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537443.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537443.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537444.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537444.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537445.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537445.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: Added label mon to host np0005537446.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: Added label _admin to host np0005537446.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:44 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:52:44 localhost ceph-mon[291772]: Deploying daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:52:44 localhost ceph-mon[291772]: mon.np0005537446@-1(synchronizing).paxosservice(auth 1..34) refresh upgraded, format 0 -> 3 Nov 27 04:52:44 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff551e0 mon_map magic: 0 from mon.1 v2:172.18.0.105:3300/0 Nov 27 04:52:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:52:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:52:44 localhost systemd[1]: tmp-crun.WlEVmu.mount: Deactivated successfully. Nov 27 04:52:44 localhost podman[291812]: 2025-11-27 09:52:44.749865927 +0000 UTC m=+0.105066216 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=Red Hat, Inc., distribution-scope=public, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, io.openshift.expose-services=, vcs-type=git, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, build-date=2025-08-20T13:12:41, name=ubi9-minimal, release=1755695350, com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, config_id=edpm, container_name=openstack_network_exporter) Nov 27 04:52:44 localhost podman[291812]: 2025-11-27 09:52:44.79354562 +0000 UTC m=+0.148745859 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, version=9.6, build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, maintainer=Red Hat, Inc., name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=openstack_network_exporter, io.buildah.version=1.33.7, vendor=Red Hat, Inc., vcs-type=git, release=1755695350, distribution-scope=public, io.openshift.expose-services=, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:52:44 localhost podman[291811]: 2025-11-27 09:52:44.790283543 +0000 UTC m=+0.145036091 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, io.buildah.version=1.41.3, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS) Nov 27 04:52:44 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:52:44 localhost podman[291811]: 2025-11-27 09:52:44.876320246 +0000 UTC m=+0.231072804 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, container_name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 04:52:44 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:52:46 localhost ceph-mon[291772]: mon.np0005537446@-1(probing) e4 my rank is now 3 (was -1) Nov 27 04:52:46 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:52:46 localhost ceph-mon[291772]: paxos.3).electionLogic(0) init, first boot, initializing epoch at 1 Nov 27 04:52:46 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e4 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:46 localhost nova_compute[284026]: 2025-11-27 09:52:46.419 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:46 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e4 adding peer [v2:172.18.0.107:3300/0,v1:172.18.0.107:6789/0] to list of hints Nov 27 04:52:46 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e4 adding peer [v2:172.18.0.107:3300/0,v1:172.18.0.107:6789/0] to list of hints Nov 27 04:52:48 localhost nova_compute[284026]: 2025-11-27 09:52:48.370 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:52:48 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e4 adding peer [v2:172.18.0.107:3300/0,v1:172.18.0.107:6789/0] to list of hints Nov 27 04:52:48 localhost podman[291850]: 2025-11-27 09:52:48.996511523 +0000 UTC m=+0.083522556 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:52:49 localhost podman[291850]: 2025-11-27 09:52:49.009831651 +0000 UTC m=+0.096842664 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:52:49 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e4 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e4 _apply_compatset_features enabling new quorum features: compat={},rocompat={},incompat={4=support erasure code pools,5=new-style osdmap encoding,6=support isa/lrc erasure code,7=support shec erasure code} Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e4 _apply_compatset_features enabling new quorum features: compat={},rocompat={},incompat={8=support monmap features,9=luminous ondisk layout,10=mimic ondisk layout,11=nautilus ondisk layout,12=octopus ondisk layout,13=pacific ondisk layout,14=quincy ondisk layout,15=reef ondisk layout} Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e4 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:49 localhost ceph-mon[291772]: mgrc update_daemon_metadata mon.np0005537446 metadata {addrs=[v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0],arch=x86_64,ceph_release=reef,ceph_version=ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable),ceph_version_short=18.2.1-361.el9cp,compression_algorithms=none, snappy, zlib, zstd, lz4,container_hostname=np0005537446.localdomain,container_image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest,cpu=AMD EPYC-Rome Processor,device_ids=,device_paths=vda=/dev/disk/by-path/pci-0000:00:04.0,devices=vda,distro=rhel,distro_description=Red Hat Enterprise Linux 9.6 (Plow),distro_version=9.6,hostname=np0005537446.localdomain,kernel_description=#1 SMP PREEMPT_DYNAMIC Wed Apr 12 10:45:03 EDT 2023,kernel_version=5.14.0-284.11.1.el9_2.x86_64,mem_swap_kb=1048572,mem_total_kb=16116612,os=Linux} Nov 27 04:52:49 localhost ceph-mon[291772]: Deploying daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537441 calling monitor election Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:52:49 localhost ceph-mon[291772]: mon.np0005537441 is new leader, mons np0005537441,np0005537443,np0005537442,np0005537446 in quorum (ranks 0,1,2,3) Nov 27 04:52:49 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:52:49 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:49 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:50 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e4 handle_auth_request failed to assign global_id Nov 27 04:52:50 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:50 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:52:50 localhost ceph-mon[291772]: Deploying daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:52:50 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e4 adding peer [v2:172.18.0.107:3300/0,v1:172.18.0.107:6789/0] to list of hints Nov 27 04:52:50 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff54f20 mon_map magic: 0 from mon.1 v2:172.18.0.105:3300/0 Nov 27 04:52:50 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:52:50 localhost ceph-mon[291772]: paxos.3).electionLogic(18) init, last seen epoch 18 Nov 27 04:52:50 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:51 localhost nova_compute[284026]: 2025-11-27 09:52:51.421 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:52 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 adding peer [v2:172.18.0.106:3300/0,v1:172.18.0.106:6789/0] to list of hints Nov 27 04:52:52 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 adding peer [v2:172.18.0.106:3300/0,v1:172.18.0.106:6789/0] to list of hints Nov 27 04:52:52 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 adding peer [v2:172.18.0.106:3300/0,v1:172.18.0.106:6789/0] to list of hints Nov 27 04:52:53 localhost nova_compute[284026]: 2025-11-27 09:52:53.373 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:54 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 adding peer [v2:172.18.0.106:3300/0,v1:172.18.0.106:6789/0] to list of hints Nov 27 04:52:55 localhost openstack_network_exporter[244641]: ERROR 09:52:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:52:55 localhost openstack_network_exporter[244641]: ERROR 09:52:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:52:55 localhost openstack_network_exporter[244641]: ERROR 09:52:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:52:55 localhost openstack_network_exporter[244641]: ERROR 09:52:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:52:55 localhost openstack_network_exporter[244641]: Nov 27 04:52:55 localhost openstack_network_exporter[244641]: ERROR 09:52:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:52:55 localhost openstack_network_exporter[244641]: Nov 27 04:52:55 localhost ceph-mon[291772]: paxos.3).electionLogic(19) init, last seen epoch 19, mid-election, bumping Nov 27 04:52:55 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:55 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e5 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e5 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537441 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537441 is new leader, mons np0005537441,np0005537443,np0005537442 in quorum (ranks 0,1,2) Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537441 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537441 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537441 is new leader, mons np0005537441,np0005537443,np0005537442,np0005537446,np0005537445 in quorum (ranks 0,1,2,3,4) Nov 27 04:52:56 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:52:56 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:52:56 localhost nova_compute[284026]: 2025-11-27 09:52:56.423 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e5 adding peer [v2:172.18.0.106:3300/0,v1:172.18.0.106:6789/0] to list of hints Nov 27 04:52:56 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff54f20 mon_map magic: 0 from mon.1 v2:172.18.0.105:3300/0 Nov 27 04:52:56 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:52:56 localhost ceph-mon[291772]: paxos.3).electionLogic(28) init, last seen epoch 28 Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e6 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:56 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e6 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:52:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:52:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:52:56 localhost podman[291928]: 2025-11-27 09:52:56.962868247 +0000 UTC m=+0.080219697 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:52:56 localhost podman[291928]: 2025-11-27 09:52:56.974028487 +0000 UTC m=+0.091379907 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:52:56 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:52:57 localhost systemd[1]: tmp-crun.gnuQWh.mount: Deactivated successfully. Nov 27 04:52:57 localhost podman[291927]: 2025-11-27 09:52:57.02842221 +0000 UTC m=+0.146803049 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, config_id=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, container_name=ovn_controller) Nov 27 04:52:57 localhost podman[291927]: 2025-11-27 09:52:57.096945761 +0000 UTC m=+0.215326640 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, config_id=ovn_controller) Nov 27 04:52:57 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:52:57 localhost podman[292049]: 2025-11-27 09:52:57.834809758 +0000 UTC m=+0.090444973 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-type=git, version=7, architecture=x86_64, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , release=553, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, build-date=2025-09-24T08:57:55, RELEASE=main) Nov 27 04:52:57 localhost podman[292049]: 2025-11-27 09:52:57.938106815 +0000 UTC m=+0.193742050 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, RELEASE=main, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, name=rhceph, release=553, io.buildah.version=1.33.12, GIT_BRANCH=main, GIT_CLEAN=True, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., distribution-scope=public, maintainer=Guillaume Abrioux , io.openshift.expose-services=, com.redhat.component=rhceph-container) Nov 27 04:52:58 localhost nova_compute[284026]: 2025-11-27 09:52:58.380 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:01 localhost nova_compute[284026]: 2025-11-27 09:53:01.425 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537446@3(electing) e6 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e6 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537441 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:53:01 localhost ceph-mon[291772]: mon.np0005537441 is new leader, mons np0005537441,np0005537443,np0005537442,np0005537446,np0005537445,np0005537444 in quorum (ranks 0,1,2,3,4,5) Nov 27 04:53:01 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:53:01 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:01 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:53:02 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:53:02 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:02 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:02 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:02 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:02 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:03 localhost nova_compute[284026]: 2025-11-27 09:53:03.380 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:04 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:05 localhost ceph-mon[291772]: Reconfiguring mon.np0005537441 (monmap changed)... Nov 27 04:53:05 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537441 on np0005537441.localdomain Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537441.xbypnw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:05 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537441.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:06 localhost nova_compute[284026]: 2025-11-27 09:53:06.428 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537441.xbypnw (monmap changed)... Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537441.xbypnw on np0005537441.localdomain Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring crash.np0005537441 (monmap changed)... Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537441 on np0005537441.localdomain Nov 27 04:53:06 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:06 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring crash.np0005537442 (monmap changed)... Nov 27 04:53:06 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537442.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:06 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537442 on np0005537442.localdomain Nov 27 04:53:06 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:07 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:07 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:07 localhost ceph-mon[291772]: Reconfiguring mon.np0005537442 (monmap changed)... Nov 27 04:53:07 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:07 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537442 on np0005537442.localdomain Nov 27 04:53:07 localhost nova_compute[284026]: 2025-11-27 09:53:07.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:08 localhost nova_compute[284026]: 2025-11-27 09:53:08.413 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:08 localhost nova_compute[284026]: 2025-11-27 09:53:08.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:08 localhost podman[242678]: time="2025-11-27T09:53:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:53:08 localhost podman[242678]: @ - - [27/Nov/2025:09:53:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:53:08 localhost podman[242678]: @ - - [27/Nov/2025:09:53:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19682 "" "Go-http-client/1.1" Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:09 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537442.byetac (monmap changed)... Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537442.byetac", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:09 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537442.byetac on np0005537442.localdomain Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' Nov 27 04:53:09 localhost ceph-mon[291772]: from='mgr.14120 172.18.0.103:0/2296951103' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:09 localhost nova_compute[284026]: 2025-11-27 09:53:09.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:09 localhost nova_compute[284026]: 2025-11-27 09:53:09.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:09 localhost nova_compute[284026]: 2025-11-27 09:53:09.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:09 localhost nova_compute[284026]: 2025-11-27 09:53:09.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:53:10 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e82 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 Nov 27 04:53:10 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e82 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 Nov 27 04:53:10 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e83 e83: 6 total, 6 up, 6 in Nov 27 04:53:10 localhost systemd[1]: session-21.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-18.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-14.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-20.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-26.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-26.scope: Consumed 3min 31.894s CPU time. Nov 27 04:53:10 localhost systemd[1]: session-16.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-23.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-24.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-22.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-17.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-19.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd[1]: session-25.scope: Deactivated successfully. Nov 27 04:53:10 localhost systemd-logind[761]: Session 25 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 19 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 17 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 24 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 26 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 18 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 14 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 20 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 21 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 16 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 23 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Session 22 logged out. Waiting for processes to exit. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 21. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 18. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 14. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 20. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 26. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 16. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 23. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 24. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 22. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 17. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 19. Nov 27 04:53:10 localhost systemd-logind[761]: Removed session 25. Nov 27 04:53:10 localhost sshd[292505]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:53:10 localhost ceph-mon[291772]: Reconfiguring mon.np0005537443 (monmap changed)... Nov 27 04:53:10 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537443 on np0005537443.localdomain Nov 27 04:53:10 localhost ceph-mon[291772]: from='client.? 172.18.0.103:0/3190559250' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:53:10 localhost ceph-mon[291772]: Activating manager daemon np0005537443.fgpbcx Nov 27 04:53:10 localhost ceph-mon[291772]: from='client.? 172.18.0.103:0/3190559250' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:53:10 localhost ceph-mon[291772]: Manager daemon np0005537443.fgpbcx is now available Nov 27 04:53:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537443.fgpbcx/mirror_snapshot_schedule"} : dispatch Nov 27 04:53:10 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537443.fgpbcx/mirror_snapshot_schedule"} : dispatch Nov 27 04:53:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537443.fgpbcx/trash_purge_schedule"} : dispatch Nov 27 04:53:10 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537443.fgpbcx/trash_purge_schedule"} : dispatch Nov 27 04:53:10 localhost systemd-logind[761]: New session 65 of user ceph-admin. Nov 27 04:53:10 localhost systemd[1]: Started Session 65 of User ceph-admin. Nov 27 04:53:10 localhost nova_compute[284026]: 2025-11-27 09:53:10.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.430 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:11 localhost podman[292591]: 2025-11-27 09:53:11.5012809 +0000 UTC m=+0.098461888 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 04:53:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:53:11 localhost podman[292591]: 2025-11-27 09:53:11.518066321 +0000 UTC m=+0.115247269 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, managed_by=edpm_ansible) Nov 27 04:53:11 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:53:11 localhost systemd[1]: tmp-crun.9OGQWm.mount: Deactivated successfully. Nov 27 04:53:11 localhost podman[292633]: 2025-11-27 09:53:11.617164235 +0000 UTC m=+0.091429139 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, config_id=multipathd, managed_by=edpm_ansible) Nov 27 04:53:11 localhost podman[292633]: 2025-11-27 09:53:11.657220872 +0000 UTC m=+0.131485826 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible) Nov 27 04:53:11 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:53:11 localhost podman[292643]: 2025-11-27 09:53:11.690462496 +0000 UTC m=+0.116766570 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, io.openshift.expose-services=, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, vendor=Red Hat, Inc., ceph=True, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, release=553, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.752 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.752 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.775 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.775 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.776 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.776 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:53:11 localhost nova_compute[284026]: 2025-11-27 09:53:11.777 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:53:11 localhost podman[292643]: 2025-11-27 09:53:11.798037438 +0000 UTC m=+0.224341522 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, RELEASE=main, release=553, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, description=Red Hat Ceph Storage 7, distribution-scope=public, GIT_CLEAN=True, architecture=x86_64, vendor=Red Hat, Inc., GIT_BRANCH=main, name=rhceph, maintainer=Guillaume Abrioux , io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=) Nov 27 04:53:12 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:12 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:12 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:12 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:12 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e6 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:53:12 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2352995063' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.286 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.509s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.348 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.349 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.570 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.572 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11478MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.572 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.573 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.645 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.645 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.646 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:53:12 localhost nova_compute[284026]: 2025-11-27 09:53:12.700 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:53:13 localhost ceph-mon[291772]: [27/Nov/2025:09:53:11] ENGINE Bus STARTING Nov 27 04:53:13 localhost ceph-mon[291772]: [27/Nov/2025:09:53:11] ENGINE Serving on https://172.18.0.105:7150 Nov 27 04:53:13 localhost ceph-mon[291772]: [27/Nov/2025:09:53:11] ENGINE Client ('172.18.0.105', 59090) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:53:13 localhost ceph-mon[291772]: [27/Nov/2025:09:53:11] ENGINE Serving on http://172.18.0.105:8765 Nov 27 04:53:13 localhost ceph-mon[291772]: [27/Nov/2025:09:53:11] ENGINE Bus STARTED Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.172 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.472s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.179 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.197 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:53:13 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e6 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:53:13 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.106:0/2871607787' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.200 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.200 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.627s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:53:13 localhost nova_compute[284026]: 2025-11-27 09:53:13.418 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:13 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e83 _set_new_cache_sizes cache_size:1019602316 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537441", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537441", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:53:14 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:53:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:53:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:53:14 localhost podman[293063]: 2025-11-27 09:53:14.93864583 +0000 UTC m=+0.091211694 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, io.openshift.expose-services=, maintainer=Red Hat, Inc., config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.buildah.version=1.33.7, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, release=1755695350, com.redhat.component=ubi9-minimal-container, url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter, architecture=x86_64, distribution-scope=public, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 04:53:14 localhost podman[293063]: 2025-11-27 09:53:14.954947278 +0000 UTC m=+0.107513142 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, container_name=openstack_network_exporter, vendor=Red Hat, Inc., io.buildah.version=1.33.7, managed_by=edpm_ansible, maintainer=Red Hat, Inc., release=1755695350, config_id=edpm, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, version=9.6, build-date=2025-08-20T13:12:41, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:53:14 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:53:15 localhost podman[293097]: 2025-11-27 09:53:15.037754214 +0000 UTC m=+0.099928697 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0) Nov 27 04:53:15 localhost podman[293097]: 2025-11-27 09:53:15.076903296 +0000 UTC m=+0.139077799 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm) Nov 27 04:53:15 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.180 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.181 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.181 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.828 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.828 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.829 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:53:15 localhost nova_compute[284026]: 2025-11-27 09:53:15.829 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:53:16 localhost nova_compute[284026]: 2025-11-27 09:53:16.238 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:53:16 localhost nova_compute[284026]: 2025-11-27 09:53:16.257 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:53:16 localhost nova_compute[284026]: 2025-11-27 09:53:16.257 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:53:16 localhost nova_compute[284026]: 2025-11-27 09:53:16.432 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:16 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:53:16 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:53:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:16 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537441.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:17 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:18 localhost nova_compute[284026]: 2025-11-27 09:53:18.458 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:18 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:18 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #13. Immutable memtables: 0. Nov 27 04:53:18 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:18.917479) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:53:18 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 3] Flushing memtable with next log file: 13 Nov 27 04:53:18 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237198917584, "job": 3, "event": "flush_started", "num_memtables": 1, "num_entries": 10895, "num_deletes": 527, "total_data_size": 14974905, "memory_usage": 15785648, "flush_reason": "Manual Compaction"} Nov 27 04:53:18 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 3] Level-0 flush table #14: started Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237199000834, "cf_name": "default", "job": 3, "event": "table_file_creation", "file_number": 14, "file_size": 10563635, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 10900, "table_properties": {"data_size": 10511725, "index_size": 26780, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 24069, "raw_key_size": 252380, "raw_average_key_size": 26, "raw_value_size": 10350308, "raw_average_value_size": 1077, "num_data_blocks": 1005, "num_entries": 9604, "num_filter_entries": 9604, "num_deletions": 526, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237164, "oldest_key_time": 1764237164, "file_creation_time": 1764237198, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 14, "seqno_to_time_mapping": "N/A"}} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 3] Flush lasted 83448 microseconds, and 16360 cpu microseconds. Nov 27 04:53:19 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e83 _set_new_cache_sizes cache_size:1020043603 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.000920) [db/flush_job.cc:967] [default] [JOB 3] Level-0 flush table #14: 10563635 bytes OK Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.000960) [db/memtable_list.cc:519] [default] Level-0 commit table #14 started Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.006072) [db/memtable_list.cc:722] [default] Level-0 commit table #14: memtable #1 done Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.006102) EVENT_LOG_v1 {"time_micros": 1764237199006095, "job": 3, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [2, 0, 0, 0, 0, 0, 0], "immutable_memtables": 0} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.006131) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: files[2 0 0 0 0 0 0] max score 0.50 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 3] Try to delete WAL files size 14901461, prev total WAL file size 14902184, number of live WAL files 2. Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.009483) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003130303430' seq:72057594037927935, type:22 .. '7061786F73003130323932' seq:0, type:0; will stop at (end) Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 4] Compacting 2@0 files to L6, score -1.00 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 3 Base level 0, inputs: [14(10MB) 8(1887B)] Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237199009637, "job": 4, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [14, 8], "score": -1, "input_data_size": 10565522, "oldest_snapshot_seqno": -1} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 4] Generated table #15: 9081 keys, 10555542 bytes, temperature: kUnknown Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237199110931, "cf_name": "default", "job": 4, "event": "table_file_creation", "file_number": 15, "file_size": 10555542, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 10505025, "index_size": 26736, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 22725, "raw_key_size": 243756, "raw_average_key_size": 26, "raw_value_size": 10350366, "raw_average_value_size": 1139, "num_data_blocks": 1003, "num_entries": 9081, "num_filter_entries": 9081, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237199, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 15, "seqno_to_time_mapping": "N/A"}} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.111353) [db/compaction/compaction_job.cc:1663] [default] [JOB 4] Compacted 2@0 files to L6 => 10555542 bytes Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.112770) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 104.2 rd, 104.1 wr, level 6, files in(2, 0) out(1 +0 blob) MB in(10.1, 0.0 +0.0 blob) out(10.1 +0.0 blob), read-write-amplify(2.0) write-amplify(1.0) OK, records in: 9609, records dropped: 528 output_compression: NoCompression Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.112803) EVENT_LOG_v1 {"time_micros": 1764237199112788, "job": 4, "event": "compaction_finished", "compaction_time_micros": 101437, "compaction_time_cpu_micros": 35230, "output_level": 6, "num_output_files": 1, "total_output_size": 10555542, "num_input_records": 9609, "num_output_records": 9081, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000014.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237199114621, "job": 4, "event": "table_file_deletion", "file_number": 14} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000008.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237199114670, "job": 4, "event": "table_file_deletion", "file_number": 8} Nov 27 04:53:19 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:53:19.009338) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:53:19 localhost ceph-mon[291772]: Reconfiguring mon.np0005537443 (monmap changed)... Nov 27 04:53:19 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:19 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537443 on np0005537443.localdomain Nov 27 04:53:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:53:19 localhost podman[293637]: 2025-11-27 09:53:19.998430957 +0000 UTC m=+0.089384774 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:53:20 localhost podman[293637]: 2025-11-27 09:53:20.038109943 +0000 UTC m=+0.129063730 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:53:20 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:53:21 localhost nova_compute[284026]: 2025-11-27 09:53:21.435 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:22 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:22 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:22 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537443.fgpbcx (monmap changed)... Nov 27 04:53:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:22 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:22 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537443.fgpbcx on np0005537443.localdomain Nov 27 04:53:22 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:23 localhost nova_compute[284026]: 2025-11-27 09:53:23.462 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:24 localhost ceph-mon[291772]: mon.np0005537446@3(peon).osd e83 _set_new_cache_sizes cache_size:1020054543 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:24 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:24 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:24 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:24 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:24 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:53:25 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:25 localhost ceph-mon[291772]: from='mgr.14184 ' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:25 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:53:25 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:53:25 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:53:25 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff55080 mon_map magic: 0 from mon.1 v2:172.18.0.105:3300/0 Nov 27 04:53:25 localhost ceph-mon[291772]: mon.np0005537446@3(peon) e7 my rank is now 2 (was 3) Nov 27 04:53:25 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.105:3300/0 Nov 27 04:53:25 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.105:3300/0 Nov 27 04:53:25 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556da996e000 mon_map magic: 0 from mon.2 v2:172.18.0.108:3300/0 Nov 27 04:53:25 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:53:25 localhost ceph-mon[291772]: paxos.2).electionLogic(32) init, last seen epoch 32 Nov 27 04:53:25 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:25 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:25 localhost openstack_network_exporter[244641]: ERROR 09:53:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:53:25 localhost openstack_network_exporter[244641]: ERROR 09:53:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:53:25 localhost openstack_network_exporter[244641]: ERROR 09:53:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:53:25 localhost openstack_network_exporter[244641]: ERROR 09:53:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:53:25 localhost openstack_network_exporter[244641]: Nov 27 04:53:25 localhost openstack_network_exporter[244641]: ERROR 09:53:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:53:25 localhost openstack_network_exporter[244641]: Nov 27 04:53:25 localhost ceph-mds[289027]: --2- [v2:172.18.0.108:6808/2569123203,v1:172.18.0.108:6809/2569123203] >> [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] conn(0x558226398400 0x55822555b180 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=1 rev1=1 crypto rx=0 tx=0 comp rx=0 tx=0).send_auth_request get_initial_auth_request returned -2 Nov 27 04:53:26 localhost nova_compute[284026]: 2025-11-27 09:53:26.437 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:53:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:53:27 localhost systemd[1]: tmp-crun.UMiL5Y.mount: Deactivated successfully. Nov 27 04:53:28 localhost podman[293663]: 2025-11-27 09:53:28.029626075 +0000 UTC m=+0.113412839 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:53:28 localhost podman[293662]: 2025-11-27 09:53:28.003573375 +0000 UTC m=+0.095444107 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, config_id=ovn_controller, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 04:53:28 localhost podman[293663]: 2025-11-27 09:53:28.066051834 +0000 UTC m=+0.149838628 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:53:28 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:53:28 localhost podman[293662]: 2025-11-27 09:53:28.087964574 +0000 UTC m=+0.179835296 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller) Nov 27 04:53:28 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:53:28 localhost nova_compute[284026]: 2025-11-27 09:53:28.483 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:30 localhost ceph-mds[289027]: mds.beacon.mds.np0005537446.hkzfou missed beacon ack from the monitors Nov 27 04:53:30 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:30 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:30 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e7 handle_timecheck drop unexpected msg Nov 27 04:53:30 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:30 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e7 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:53:31 localhost nova_compute[284026]: 2025-11-27 09:53:31.439 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:53:31 localhost ceph-mon[291772]: Health check failed: 1/5 mons down, quorum np0005537443,np0005537446,np0005537445,np0005537444 (MON_DOWN) Nov 27 04:53:31 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:53:31 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537442,np0005537446,np0005537445,np0005537444 in quorum (ranks 0,1,2,3,4) Nov 27 04:53:31 localhost ceph-mon[291772]: Health check cleared: MON_DOWN (was: 1/5 mons down, quorum np0005537443,np0005537446,np0005537445,np0005537444) Nov 27 04:53:31 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:53:31 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:31 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:53:31 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:31 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:32 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:53:32 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:53:32 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:32 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:32 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:33 localhost nova_compute[284026]: 2025-11-27 09:53:33.489 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:33 localhost ceph-mon[291772]: Removed label mon from host np0005537441.localdomain Nov 27 04:53:33 localhost ceph-mon[291772]: Reconfiguring mon.np0005537444 (monmap changed)... Nov 27 04:53:33 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:33 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:53:33 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:33 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:33 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:33 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:34 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054729 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:34 localhost ceph-mon[291772]: Removed label mgr from host np0005537441.localdomain Nov 27 04:53:34 localhost ceph-mon[291772]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:53:34 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:53:34 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:34 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:34 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:53:35 localhost ceph-mon[291772]: Reconfiguring osd.1 (monmap changed)... Nov 27 04:53:35 localhost ceph-mon[291772]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:53:35 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:35 localhost ceph-mon[291772]: Removed label _admin from host np0005537441.localdomain Nov 27 04:53:35 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:35 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:35 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:53:36 localhost nova_compute[284026]: 2025-11-27 09:53:36.441 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:36 localhost ceph-mon[291772]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:53:36 localhost ceph-mon[291772]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:53:36 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:36 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:36 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:53:37 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:53:37 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:53:37 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:37 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:37 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:38 localhost nova_compute[284026]: 2025-11-27 09:53:38.533 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:38 localhost podman[242678]: time="2025-11-27T09:53:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:53:38 localhost podman[242678]: @ - - [27/Nov/2025:09:53:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:53:38 localhost podman[242678]: @ - - [27/Nov/2025:09:53:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19693 "" "Go-http-client/1.1" Nov 27 04:53:38 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:53:38 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:53:38 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:38 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:38 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:39 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:39 localhost podman[293764]: Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.630837827 +0000 UTC m=+0.084916753 container create cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, maintainer=Guillaume Abrioux , ceph=True, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, CEPH_POINT_RELEASE=, architecture=x86_64, vendor=Red Hat, Inc., release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, version=7, GIT_BRANCH=main, RELEASE=main, com.redhat.component=rhceph-container, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:53:39 localhost systemd[1]: Started libpod-conmon-cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293.scope. Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.591928521 +0000 UTC m=+0.046007457 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:39 localhost systemd[1]: Started libcrun container. Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.712136513 +0000 UTC m=+0.166215439 container init cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, version=7, release=553, GIT_BRANCH=main, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, io.openshift.expose-services=, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, ceph=True, GIT_CLEAN=True, architecture=x86_64, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.724172857 +0000 UTC m=+0.178251783 container start cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, vcs-type=git, release=553, RELEASE=main, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux ) Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.724426334 +0000 UTC m=+0.178505320 container attach cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, GIT_CLEAN=True, version=7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., release=553, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, ceph=True, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , RELEASE=main, io.buildah.version=1.33.12, vcs-type=git) Nov 27 04:53:39 localhost stoic_bhabha[293780]: 167 167 Nov 27 04:53:39 localhost systemd[1]: libpod-cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293.scope: Deactivated successfully. Nov 27 04:53:39 localhost podman[293764]: 2025-11-27 09:53:39.731933495 +0000 UTC m=+0.186012451 container died cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, ceph=True, description=Red Hat Ceph Storage 7, name=rhceph, architecture=x86_64, distribution-scope=public, GIT_BRANCH=main, RELEASE=main, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, GIT_CLEAN=True, release=553, version=7) Nov 27 04:53:39 localhost podman[293785]: 2025-11-27 09:53:39.873116321 +0000 UTC m=+0.126955194 container remove cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stoic_bhabha, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, version=7, release=553, distribution-scope=public, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, architecture=x86_64, GIT_CLEAN=True, GIT_BRANCH=main, ceph=True, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph) Nov 27 04:53:39 localhost systemd[1]: libpod-conmon-cc4a500b61598a62fe28670f91c902a4072128ceb16ff7703664f92c6c14d293.scope: Deactivated successfully. Nov 27 04:53:40 localhost ceph-mon[291772]: Reconfiguring mon.np0005537445 (monmap changed)... Nov 27 04:53:40 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:40 localhost ceph-mon[291772]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:40 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:40 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:53:40 localhost podman[293853]: Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.611141142 +0000 UTC m=+0.076982460 container create 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, architecture=x86_64, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, vendor=Red Hat, Inc., io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, ceph=True, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux ) Nov 27 04:53:40 localhost systemd[1]: var-lib-containers-storage-overlay-02772e1ce2f069bfde3791d8ba3393d2d279243419a0cc396169b36ada7d18c9-merged.mount: Deactivated successfully. Nov 27 04:53:40 localhost systemd[1]: Started libpod-conmon-1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa.scope. Nov 27 04:53:40 localhost systemd[1]: Started libcrun container. Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.581651289 +0000 UTC m=+0.047492617 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.687210507 +0000 UTC m=+0.153051845 container init 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, distribution-scope=public, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, version=7, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, vcs-type=git, ceph=True) Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.698440669 +0000 UTC m=+0.164281997 container start 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, GIT_BRANCH=main, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, vendor=Red Hat, Inc., release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, version=7, CEPH_POINT_RELEASE=, distribution-scope=public, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True) Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.699250411 +0000 UTC m=+0.165091739 container attach 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, vendor=Red Hat, Inc., version=7, GIT_BRANCH=main, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, distribution-scope=public, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, build-date=2025-09-24T08:57:55, RELEASE=main) Nov 27 04:53:40 localhost strange_hoover[293868]: 167 167 Nov 27 04:53:40 localhost systemd[1]: libpod-1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa.scope: Deactivated successfully. Nov 27 04:53:40 localhost podman[293853]: 2025-11-27 09:53:40.70220521 +0000 UTC m=+0.168046518 container died 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, name=rhceph, CEPH_POINT_RELEASE=, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , version=7, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, vendor=Red Hat, Inc., RELEASE=main, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, release=553, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, architecture=x86_64) Nov 27 04:53:40 localhost podman[293873]: 2025-11-27 09:53:40.795021666 +0000 UTC m=+0.079812238 container remove 1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=strange_hoover, architecture=x86_64, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, io.openshift.expose-services=, io.buildah.version=1.33.12, ceph=True, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, release=553, distribution-scope=public, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:53:40 localhost systemd[1]: libpod-conmon-1cdcbc75d174b792089a65bdd2bde82a5807254803153b43d095e8c4fcd88cfa.scope: Deactivated successfully. Nov 27 04:53:41 localhost ceph-mon[291772]: Reconfiguring osd.0 (monmap changed)... Nov 27 04:53:41 localhost ceph-mon[291772]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:53:41 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:41 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:41 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch Nov 27 04:53:41 localhost nova_compute[284026]: 2025-11-27 09:53:41.443 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:41 localhost podman[293950]: Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.601155998 +0000 UTC m=+0.078209154 container create da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, architecture=x86_64, description=Red Hat Ceph Storage 7, distribution-scope=public, vendor=Red Hat, Inc., ceph=True, GIT_BRANCH=main, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , io.openshift.expose-services=) Nov 27 04:53:41 localhost systemd[1]: Started libpod-conmon-da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2.scope. Nov 27 04:53:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:53:41 localhost systemd[1]: var-lib-containers-storage-overlay-340cdd355ac0a901ac37717366ca5eccda8d5e8fb5169aa6c369dc3758a913aa-merged.mount: Deactivated successfully. Nov 27 04:53:41 localhost systemd[1]: Started libcrun container. Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.666069113 +0000 UTC m=+0.143122259 container init da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, ceph=True, RELEASE=main, vcs-type=git, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, GIT_BRANCH=main, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, distribution-scope=public, release=553, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.567894213 +0000 UTC m=+0.044947399 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:41 localhost vigorous_jepsen[293965]: 167 167 Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.679607727 +0000 UTC m=+0.156660873 container start da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, GIT_CLEAN=True, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, build-date=2025-09-24T08:57:55, RELEASE=main, ceph=True, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, version=7, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, io.openshift.expose-services=, CEPH_POINT_RELEASE=) Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.679940246 +0000 UTC m=+0.156993412 container attach da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, architecture=x86_64, name=rhceph, GIT_BRANCH=main, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, RELEASE=main, io.buildah.version=1.33.12, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., vcs-type=git) Nov 27 04:53:41 localhost podman[293950]: 2025-11-27 09:53:41.684415615 +0000 UTC m=+0.161468791 container died da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, name=rhceph, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, version=7, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., GIT_CLEAN=True, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph) Nov 27 04:53:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:53:41 localhost systemd[1]: libpod-da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2.scope: Deactivated successfully. Nov 27 04:53:41 localhost podman[293967]: 2025-11-27 09:53:41.754240293 +0000 UTC m=+0.097386389 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:53:41 localhost podman[293979]: 2025-11-27 09:53:41.785660888 +0000 UTC m=+0.091834630 container remove da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_jepsen, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, ceph=True, version=7, io.openshift.expose-services=, CEPH_POINT_RELEASE=, distribution-scope=public, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, RELEASE=main, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, architecture=x86_64, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc.) Nov 27 04:53:41 localhost systemd[1]: libpod-conmon-da542524db068d149fadcd03c263a2a74cbf3c69326890e1ed4c6614586222a2.scope: Deactivated successfully. Nov 27 04:53:41 localhost podman[293967]: 2025-11-27 09:53:41.795237035 +0000 UTC m=+0.138383131 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 04:53:41 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:53:41 localhost podman[293981]: 2025-11-27 09:53:41.852696489 +0000 UTC m=+0.150510896 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, config_id=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, container_name=multipathd) Nov 27 04:53:41 localhost podman[293981]: 2025-11-27 09:53:41.87504701 +0000 UTC m=+0.172861417 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:53:41 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:53:42 localhost ceph-mon[291772]: Reconfiguring osd.3 (monmap changed)... Nov 27 04:53:42 localhost ceph-mon[291772]: Reconfiguring daemon osd.3 on np0005537446.localdomain Nov 27 04:53:42 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:42 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:42 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:53:42 localhost systemd[1]: var-lib-containers-storage-overlay-fd3becdced1bd0c8c70900df9f9992ad3bb686ac1e2753d0077520316c2a163c-merged.mount: Deactivated successfully. Nov 27 04:53:42 localhost podman[294081]: Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.692455516 +0000 UTC m=+0.080954768 container create 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, io.openshift.expose-services=, architecture=x86_64, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, release=553, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, ceph=True, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, version=7) Nov 27 04:53:42 localhost systemd[1]: Started libpod-conmon-15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba.scope. Nov 27 04:53:42 localhost systemd[1]: Started libcrun container. Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.662171871 +0000 UTC m=+0.050671153 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.770140144 +0000 UTC m=+0.158639406 container init 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., GIT_CLEAN=True, vcs-type=git, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, RELEASE=main, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, distribution-scope=public, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , architecture=x86_64, CEPH_POINT_RELEASE=, release=553, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.780852153 +0000 UTC m=+0.169351405 container start 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, release=553, RELEASE=main, description=Red Hat Ceph Storage 7, distribution-scope=public, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, vendor=Red Hat, Inc., version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.781208372 +0000 UTC m=+0.169707624 container attach 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, version=7, vcs-type=git, ceph=True, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, CEPH_POINT_RELEASE=, release=553, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:53:42 localhost relaxed_germain[294096]: 167 167 Nov 27 04:53:42 localhost systemd[1]: libpod-15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba.scope: Deactivated successfully. Nov 27 04:53:42 localhost podman[294081]: 2025-11-27 09:53:42.786423222 +0000 UTC m=+0.174922474 container died 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, ceph=True, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , RELEASE=main, io.openshift.expose-services=, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, vcs-type=git, GIT_CLEAN=True, io.buildah.version=1.33.12) Nov 27 04:53:42 localhost podman[294101]: 2025-11-27 09:53:42.893561882 +0000 UTC m=+0.094758788 container remove 15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=relaxed_germain, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, version=7, name=rhceph, release=553, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, io.openshift.expose-services=, ceph=True) Nov 27 04:53:42 localhost systemd[1]: libpod-conmon-15d9e7326e1ee7e80a74c6587409f7221527a6e148640ba9316dd7eb0ccef4ba.scope: Deactivated successfully. Nov 27 04:53:43 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537446.hkzfou (monmap changed)... Nov 27 04:53:43 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537446.hkzfou on np0005537446.localdomain Nov 27 04:53:43 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:43 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:43 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:43 localhost nova_compute[284026]: 2025-11-27 09:53:43.532 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:53:43.563 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:53:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:53:43.563 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:53:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:53:43.564 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:53:43 localhost podman[294170]: Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.640495963 +0000 UTC m=+0.083454544 container create b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, version=7, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., RELEASE=main, io.openshift.expose-services=, release=553, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, architecture=x86_64, ceph=True, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7) Nov 27 04:53:43 localhost systemd[1]: var-lib-containers-storage-overlay-85bde2a7461836e003a1999fd25f9c71f66aceb330fd0022db725ba789e86f01-merged.mount: Deactivated successfully. Nov 27 04:53:43 localhost systemd[1]: Started libpod-conmon-b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a.scope. Nov 27 04:53:43 localhost systemd[1]: Started libcrun container. Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.605011649 +0000 UTC m=+0.047970250 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.708105121 +0000 UTC m=+0.151063672 container init b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, distribution-scope=public, maintainer=Guillaume Abrioux , version=7, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, architecture=x86_64, vcs-type=git, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, ceph=True, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, release=553) Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.720113194 +0000 UTC m=+0.163071775 container start b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, com.redhat.component=rhceph-container, distribution-scope=public, GIT_BRANCH=main, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, name=rhceph, architecture=x86_64, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.720413652 +0000 UTC m=+0.163372203 container attach b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, GIT_CLEAN=True, ceph=True, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, release=553, RELEASE=main, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., io.buildah.version=1.33.12, name=rhceph, com.redhat.component=rhceph-container) Nov 27 04:53:43 localhost zen_taussig[294184]: 167 167 Nov 27 04:53:43 localhost systemd[1]: libpod-b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a.scope: Deactivated successfully. Nov 27 04:53:43 localhost podman[294170]: 2025-11-27 09:53:43.723007681 +0000 UTC m=+0.165966282 container died b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, RELEASE=main, com.redhat.component=rhceph-container, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, version=7, architecture=x86_64, vendor=Red Hat, Inc., io.openshift.expose-services=, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, ceph=True, distribution-scope=public, build-date=2025-09-24T08:57:55, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:53:43 localhost podman[294189]: 2025-11-27 09:53:43.82637896 +0000 UTC m=+0.090115723 container remove b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_taussig, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., release=553, vcs-type=git, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, ceph=True, version=7, GIT_CLEAN=True, name=rhceph) Nov 27 04:53:43 localhost systemd[1]: libpod-conmon-b4490acfdae674585d409e1c339ccc0aff65b63ec7dd78b7f92d931c38b05a3a.scope: Deactivated successfully. Nov 27 04:53:44 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537446.nfulyg (monmap changed)... Nov 27 04:53:44 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537446.nfulyg on np0005537446.localdomain Nov 27 04:53:44 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:44 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:44 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.136 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.138 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.170 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.171 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9cb2cbfb-dff8-43bb-b794-4ccfec26fd3c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.138725', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f5f79d22-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': 'e7a858fa6b51e71011191aa618d0cd0e6d1dd884c15cbf94659c50b555511234'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.138725', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f5f7b898-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '9ef64900ea83f5bc45f4b259b7691a0569601e73289d2c50dcf5317d3ddb15de'}]}, 'timestamp': '2025-11-27 09:53:44.171994', '_unique_id': '7814faddf0634b57995105aae07e576b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.174 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.176 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.176 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.191 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.191 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd794d3c5-a394-40e3-9820-277d932fa383', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.176571', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f5fac682-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': 'f192752a177c825edf2385e32a6252544f0aea0c75572fc2dec6f1bd05891ff1'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.176571', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f5fad8e8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': 'e393280232de68e02279898f5717d3fe3656eb9fba526c77166866e165e2a869'}]}, 'timestamp': '2025-11-27 09:53:44.192437', '_unique_id': '7ec23c83eb7c40b095220a607a66eb10'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.193 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.194 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.198 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '61f34e39-75a9-43ec-a04d-770f994ee9a1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.195001', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f5fbe63e-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': '13a2de0346701805c3015cd57e243ac8ea4cee5374f15f78cb8c7a71b8b4d53c'}]}, 'timestamp': '2025-11-27 09:53:44.199365', '_unique_id': 'dba511e167d54025901fa303f6b12401'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.200 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.201 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.201 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.202 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4b8d7f6d-511d-4f5b-bb68-5634252d38bc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.202050', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f5fc61fe-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': 'b421ff95b08e55c92048cf0e3cc6165340eee5e84c698be2ec856c516f23e90b'}]}, 'timestamp': '2025-11-27 09:53:44.202536', '_unique_id': '0b18063e59064ea1b63db7b659119055'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.204 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.204 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.222 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 12410000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '888ddb66-2a89-4b88-ae4f-a10ca6018b6e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 12410000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:53:44.204911', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'f5ff9842-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.410861292, 'message_signature': 'acbf39c5d342eaec9b83c0b091083a1aac3e05af5d13c72aa504a2278d07b49a'}]}, 'timestamp': '2025-11-27 09:53:44.223664', '_unique_id': '50655697101d4defaedfe5324eb5b70b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.226 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a94710b3-058b-4e71-a207-09779528b6dd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.226499', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f6001efc-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': '96a73b451f34305b740112f844894c34818d5add193c3beac24a753a26108c61'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.226499', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f60031b2-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': '8ccb3381c26a0cb93f575f7e9b0cebe48ae92d7f7d0200986f564c88c5d873e5'}]}, 'timestamp': '2025-11-27 09:53:44.227477', '_unique_id': '54094ee8a4ff492093c20265af5afbb2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.229 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e0fb230f-47db-4c09-9583-8a35e1bb7eef', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.230096', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f600ac32-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': 'fc33814021b09a720ad73e60c54aa233494ccf2ab5505f08daebca4db7ac9fb8'}]}, 'timestamp': '2025-11-27 09:53:44.230691', '_unique_id': 'b1fccb5261294fcab798c267a447130b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'dfbc845d-5c9d-4f40-b57f-6f3e237918b7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.233654', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f60134d6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': 'f48421afe3eba197d21b9a4bf9311d625ecaeb40bc1a44a8f73d321854820c16'}]}, 'timestamp': '2025-11-27 09:53:44.234166', '_unique_id': 'ba71d5e1ccff4cc0beaef4b1a05eaf4e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '98a94192-8e63-4d28-8859-0db3520623f0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.236582', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f601a9ac-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': 'b8a7301e95b1683d800ded0c7543cbe77be1ab0d97a399a7700f99e72b83c1fb'}]}, 'timestamp': '2025-11-27 09:53:44.237160', '_unique_id': 'adc179bd30e743388b79cc0f14676bec'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.239 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'dafd4f0d-bd6f-4981-8688-bd77a4765cde', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.239812', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f60224a4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': '74b14c57d58232c1ad33b941a9c073d796ecf273d9719fb243f2f4c54a5b605e'}]}, 'timestamp': '2025-11-27 09:53:44.240214', '_unique_id': '9168ce1f9a214f849a273adf1226055d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.241 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8e08cb1e-ad5c-472b-8daf-6649307b0762', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.241676', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f6026a4a-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '1ec265143f99fed8cb83960aab35f449572eec20f4cb932db8535d332df83b70'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.241676', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f602767a-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '1a58c7573f6c77573aa8e486481f7d3d12265ee7101782d9d9cdd52724bda240'}]}, 'timestamp': '2025-11-27 09:53:44.242268', '_unique_id': '33b64214bc2049f1a18a4088b4dabe38'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.243 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.243 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '931d2ce4-e68c-4ef7-92fb-9de533f03085', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.243780', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f602bcd4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': 'e95c69b93aa2fb20a060b0ec4114301a218921da51ca1ecac3a4180640841765'}]}, 'timestamp': '2025-11-27 09:53:44.244100', '_unique_id': 'ba4ed1d77a67402dad33c70c5b29c437'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6ce35931-1950-4da1-b420-7df3350e3df5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.245567', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f6030324-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': '90bd733003c1696e02d5aead00b4deb4459e9696031aa03ccbcf5f9f6de7ff04'}]}, 'timestamp': '2025-11-27 09:53:44.245887', '_unique_id': 'e545bb3fa3054b1c97a16bd94c3407dc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a817d184-d5c4-473e-ae4c-7683292f0da4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.247337', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f6034758-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': 'a8353691ecb1da7b3bfdf5cd3c2fb7039462de3145b4ae1cadf3b4fb69b80e13'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.247337', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f60353ce-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '84076f2dfe96de1ef41623a70c0ca87f276d445c4a5d7d6d1f726229e3307dd5'}]}, 'timestamp': '2025-11-27 09:53:44.247933', '_unique_id': '9258684ec40545e29e8e32344c226b13'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.249 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1b38523b-056a-44c9-aa74-0bc5f9accca7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.249421', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f60399d8-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '4612adbca8e01bd10cb90a521450e67c477e080366680e90e15afb76983cc7ae'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.249421', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f603a45a-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '28f4d5731da74cfecf8a77de70e60abc1160180b13b2e092dbd26cfc98668da9'}]}, 'timestamp': '2025-11-27 09:53:44.250011', '_unique_id': '8d3d50f06ba14106b1355856f85bc4e5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.251 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.251 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bce19877-d3f5-4cbc-945c-a618b7368199', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:53:44.251603', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'f603ee56-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.410861292, 'message_signature': '69c32dfa2d43fd8c710d75258c1f8fb2b3026ae6c7c9efea3a2a5702555d5b4d'}]}, 'timestamp': '2025-11-27 09:53:44.251894', '_unique_id': '7e6079623e7f4ff292337fce9171d5ff'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '01ad4556-6b78-4bfa-90a8-3684f1b4b5ee', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.253285', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f6042fba-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': '4cc777d1063393f540c2030fcac6fdf34dddd9217246fc97536823946e02653c'}]}, 'timestamp': '2025-11-27 09:53:44.253607', '_unique_id': '5074b36df52a4a78a1808999ed2b1d4e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.254 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6de29cbf-1eaa-4c37-97dd-b6389d9c15b1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:53:44.255169', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'f60479c0-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.383067105, 'message_signature': '28514f9ddc1bc54b7df8367bc1f19a2ddd03a70022a5e1a3c87d7f7db9ce3449'}]}, 'timestamp': '2025-11-27 09:53:44.255475', '_unique_id': '08df00c78f034330aff376bd90e1f540'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.257 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5e7e09e2-e945-4619-81f3-e53ad6cb58a7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.257511', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f604d5e6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': '816a95159bd940058748b38d7be2e475d51517718b191da84106c21f229c39ef'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.257511', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f604e09a-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.36467873, 'message_signature': '9cd6bf192169f1404bb03b99ff108d5058e3c19bd6d719210a2e667dc97d1a3d'}]}, 'timestamp': '2025-11-27 09:53:44.258101', '_unique_id': '4677b56b2dae4e9baed350b7c7c04664'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8ef6074d-e968-4a23-aea1-bf12e61b3ca4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.259635', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f60527c6-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '7017a9738d8b349e7ed3b0833201240c0bdc0e252523f694bcb307cfb8a34ab1'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.259635', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f6053662-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '643f4130fc5ba75bda99068dee05ed5c8a337d46db6cbded0e55ebe9d9f92326'}]}, 'timestamp': '2025-11-27 09:53:44.260297', '_unique_id': '3249b1879a254c2e964e1341d1ecfcc2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.263 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.263 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.263 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '19c1dd4c-338a-429b-9d99-354b3852eb93', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:53:44.263139', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'f605b0c4-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': 'ce074c5821c9cbe555a62e35bb34d77f03cbbd5ca740879a34c2da4adc06418d'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:53:44.263139', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'f605bc40-cb76-11f0-8166-fa163e6bfee1', 'monotonic_time': 11454.326799792, 'message_signature': '5983cb316b12385c59eaf7ac146965d5c6f4bb831659a734f465cb233888ce91'}]}, 'timestamp': '2025-11-27 09:53:44.263713', '_unique_id': '834d35975523460c8719fd6ea52981e8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:53:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:53:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:53:44 localhost podman[294260]: Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.644101574 +0000 UTC m=+0.082254492 container create 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, vcs-type=git, vendor=Red Hat, Inc., io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, release=553, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, version=7, RELEASE=main, ceph=True, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, GIT_CLEAN=True, description=Red Hat Ceph Storage 7) Nov 27 04:53:44 localhost systemd[1]: tmp-crun.CKIvrV.mount: Deactivated successfully. Nov 27 04:53:44 localhost systemd[1]: var-lib-containers-storage-overlay-aab62d46aae11d66cea53017e3a7f151b7f49458eeeb421b94d91cbac286554c-merged.mount: Deactivated successfully. Nov 27 04:53:44 localhost systemd[1]: Started libpod-conmon-3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3.scope. Nov 27 04:53:44 localhost systemd[1]: Started libcrun container. Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.61085435 +0000 UTC m=+0.049007278 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.716077419 +0000 UTC m=+0.154230337 container init 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, build-date=2025-09-24T08:57:55, name=rhceph, vcs-type=git, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , ceph=True, io.openshift.tags=rhceph ceph, architecture=x86_64, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.725502852 +0000 UTC m=+0.163655740 container start 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, description=Red Hat Ceph Storage 7, vcs-type=git, build-date=2025-09-24T08:57:55, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, release=553, RELEASE=main, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, ceph=True, architecture=x86_64, GIT_BRANCH=main, io.openshift.expose-services=, distribution-scope=public, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True) Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.725647836 +0000 UTC m=+0.163800744 container attach 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, com.redhat.component=rhceph-container, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, version=7, distribution-scope=public, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , name=rhceph, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, GIT_BRANCH=main) Nov 27 04:53:44 localhost tender_wozniak[294275]: 167 167 Nov 27 04:53:44 localhost systemd[1]: libpod-3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3.scope: Deactivated successfully. Nov 27 04:53:44 localhost podman[294260]: 2025-11-27 09:53:44.73283844 +0000 UTC m=+0.170991368 container died 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, com.redhat.component=rhceph-container, ceph=True, version=7, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, RELEASE=main, distribution-scope=public, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:53:44 localhost podman[294280]: 2025-11-27 09:53:44.835643483 +0000 UTC m=+0.088913590 container remove 3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=tender_wozniak, GIT_BRANCH=main, architecture=x86_64, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, RELEASE=main, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, name=rhceph, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:53:44 localhost systemd[1]: libpod-conmon-3a111275f49c823718e35f899c3713a071e38823013e5d92500ad937fdc489a3.scope: Deactivated successfully. Nov 27 04:53:45 localhost ceph-mon[291772]: Reconfiguring mon.np0005537446 (monmap changed)... Nov 27 04:53:45 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:53:45 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:45 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:53:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:53:45 localhost systemd[1]: var-lib-containers-storage-overlay-4c0ba0f5f0c7bb8246fa3df742ade71736301d24c1d628a91fb17292d0110415-merged.mount: Deactivated successfully. Nov 27 04:53:45 localhost podman[294297]: 2025-11-27 09:53:45.751515446 +0000 UTC m=+0.093782871 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:53:45 localhost podman[294297]: 2025-11-27 09:53:45.766085158 +0000 UTC m=+0.108352613 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, container_name=ceilometer_agent_compute) Nov 27 04:53:45 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:53:45 localhost podman[294298]: 2025-11-27 09:53:45.85913921 +0000 UTC m=+0.198284482 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.buildah.version=1.33.7, managed_by=edpm_ansible, name=ubi9-minimal, release=1755695350, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, maintainer=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, architecture=x86_64, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., version=9.6, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container) Nov 27 04:53:45 localhost podman[294298]: 2025-11-27 09:53:45.877990296 +0000 UTC m=+0.217135608 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, container_name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=Red Hat, Inc., release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, distribution-scope=public, managed_by=edpm_ansible, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, config_id=edpm, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, version=9.6, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, name=ubi9-minimal) Nov 27 04:53:45 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:53:46 localhost nova_compute[284026]: 2025-11-27 09:53:46.446 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:47 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: Added label _no_schedule to host np0005537441.localdomain Nov 27 04:53:48 localhost ceph-mon[291772]: Added label SpecialHostLabels.DRAIN_CONF_KEYRING to host np0005537441.localdomain Nov 27 04:53:48 localhost ceph-mon[291772]: Removing np0005537441.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Removing np0005537441.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:53:48 localhost ceph-mon[291772]: Removing np0005537441.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:48 localhost nova_compute[284026]: 2025-11-27 09:53:48.576 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:49 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:49 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:49 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:53:49 localhost ceph-mon[291772]: Removing daemon crash.np0005537441 from np0005537441.localdomain -- ports [] Nov 27 04:53:49 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:53:50 localhost podman[294692]: 2025-11-27 09:53:50.535700594 +0000 UTC m=+0.080613588 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain"} : dispatch Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain"}]': finished Nov 27 04:53:50 localhost ceph-mon[291772]: Removing key for client.crash.np0005537441.localdomain Nov 27 04:53:50 localhost ceph-mon[291772]: Removed host np0005537441.localdomain Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth rm", "entity": "client.crash.np0005537441.localdomain"} : dispatch Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd='[{"prefix": "auth rm", "entity": "client.crash.np0005537441.localdomain"}]': finished Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:50 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537442.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:50 localhost podman[294692]: 2025-11-27 09:53:50.571479266 +0000 UTC m=+0.116392240 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:53:50 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:53:51 localhost nova_compute[284026]: 2025-11-27 09:53:51.447 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:52 localhost ceph-mon[291772]: Reconfiguring crash.np0005537442 (monmap changed)... Nov 27 04:53:52 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537442 on np0005537442.localdomain Nov 27 04:53:52 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:52 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:52 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:53 localhost ceph-mon[291772]: Reconfiguring mon.np0005537442 (monmap changed)... Nov 27 04:53:53 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537442 on np0005537442.localdomain Nov 27 04:53:53 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:53 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:53 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537442.byetac", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:53 localhost nova_compute[284026]: 2025-11-27 09:53:53.578 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:54 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537442.byetac (monmap changed)... Nov 27 04:53:54 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537442.byetac on np0005537442.localdomain Nov 27 04:53:54 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:54 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:54 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:53:54 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:55 localhost ceph-mon[291772]: Reconfiguring mon.np0005537443 (monmap changed)... Nov 27 04:53:55 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537443 on np0005537443.localdomain Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:55 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:55 localhost openstack_network_exporter[244641]: ERROR 09:53:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:53:55 localhost openstack_network_exporter[244641]: ERROR 09:53:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:53:55 localhost openstack_network_exporter[244641]: ERROR 09:53:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:53:55 localhost openstack_network_exporter[244641]: ERROR 09:53:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:53:55 localhost openstack_network_exporter[244641]: Nov 27 04:53:55 localhost openstack_network_exporter[244641]: ERROR 09:53:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:53:55 localhost openstack_network_exporter[244641]: Nov 27 04:53:56 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537443.fgpbcx (monmap changed)... Nov 27 04:53:56 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537443.fgpbcx on np0005537443.localdomain Nov 27 04:53:56 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:53:56 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:53:56 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:56 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:56 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:53:56 localhost nova_compute[284026]: 2025-11-27 09:53:56.448 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:57 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:53:57 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:53:57 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:57 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:57 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:53:57 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:53:57 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:53:58 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:53:58 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:58 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:58 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:53:58 localhost ceph-mon[291772]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:53:58 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:53:58 localhost ceph-mon[291772]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:53:58 localhost nova_compute[284026]: 2025-11-27 09:53:58.613 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:53:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:53:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:53:58 localhost systemd[1]: tmp-crun.1ja4b8.mount: Deactivated successfully. Nov 27 04:53:58 localhost podman[294717]: 2025-11-27 09:53:58.997753448 +0000 UTC m=+0.090927936 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible) Nov 27 04:53:59 localhost podman[294717]: 2025-11-27 09:53:59.039892081 +0000 UTC m=+0.133066539 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.build-date=20251125) Nov 27 04:53:59 localhost podman[294718]: 2025-11-27 09:53:59.055403907 +0000 UTC m=+0.145903623 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:53:59 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:53:59 localhost podman[294718]: 2025-11-27 09:53:59.0673948 +0000 UTC m=+0.157894556 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:53:59 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:53:59 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:53:59 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556d9ff55080 mon_map magic: 0 from mon.2 v2:172.18.0.108:3300/0 Nov 27 04:53:59 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:53:59 localhost ceph-mon[291772]: paxos.2).electionLogic(38) init, last seen epoch 38 Nov 27 04:53:59 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e8 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:01 localhost nova_compute[284026]: 2025-11-27 09:54:01.469 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:03 localhost nova_compute[284026]: 2025-11-27 09:54:03.615 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:04 localhost ceph-mon[291772]: paxos.2).electionLogic(39) init, last seen epoch 39, mid-election, bumping Nov 27 04:54:04 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e8 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:04 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e8 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:04 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e8 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:04 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:54:05 localhost ceph-mon[291772]: Health check failed: 1/4 mons down, quorum np0005537443,np0005537442,np0005537445 (MON_DOWN) Nov 27 04:54:05 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:54:05 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537442,np0005537446,np0005537445 in quorum (ranks 0,1,2,3) Nov 27 04:54:05 localhost ceph-mon[291772]: Health check cleared: MON_DOWN (was: 1/4 mons down, quorum np0005537443,np0005537442,np0005537445) Nov 27 04:54:05 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:54:05 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:54:05 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:05 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:05 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:05 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:54:05 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:05 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:05 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:05 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:05 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:06 localhost nova_compute[284026]: 2025-11-27 09:54:06.471 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:06 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:06 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:06 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:06 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:06 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:06 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537442.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:08 localhost ceph-mon[291772]: Reconfiguring crash.np0005537442 (monmap changed)... Nov 27 04:54:08 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537442 on np0005537442.localdomain Nov 27 04:54:08 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:08 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:08 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537442.byetac", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:08 localhost nova_compute[284026]: 2025-11-27 09:54:08.674 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:08 localhost podman[242678]: time="2025-11-27T09:54:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:54:08 localhost podman[242678]: @ - - [27/Nov/2025:09:54:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:54:08 localhost podman[242678]: @ - - [27/Nov/2025:09:54:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19695 "" "Go-http-client/1.1" Nov 27 04:54:09 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537442.byetac (monmap changed)... Nov 27 04:54:09 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537442.byetac on np0005537442.localdomain Nov 27 04:54:09 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:09 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:09 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:09 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:09 localhost nova_compute[284026]: 2025-11-27 09:54:09.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:09 localhost nova_compute[284026]: 2025-11-27 09:54:09.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:10 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537443.fgpbcx (monmap changed)... Nov 27 04:54:10 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537443.fgpbcx on np0005537443.localdomain Nov 27 04:54:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:10 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:10 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:54:10 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.107:0/1547453768' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:54:10 localhost nova_compute[284026]: 2025-11-27 09:54:10.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:11 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:54:11 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:54:11 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:11 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:11 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.513 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:11 localhost nova_compute[284026]: 2025-11-27 09:54:11.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:54:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:54:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:54:12 localhost podman[295122]: 2025-11-27 09:54:12.009176078 +0000 UTC m=+0.090142174 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_metadata_agent, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 04:54:12 localhost podman[295122]: 2025-11-27 09:54:12.047211131 +0000 UTC m=+0.128177247 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:54:12 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:54:12 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:54:12 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:54:12 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:12 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:12 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:54:12 localhost podman[295123]: 2025-11-27 09:54:12.055102533 +0000 UTC m=+0.129615396 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 04:54:12 localhost podman[295123]: 2025-11-27 09:54:12.142926554 +0000 UTC m=+0.217439407 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:54:12 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:54:13 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:54:13 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:54:13 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:13 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:13 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:54:13 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:13 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.712 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.752 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.752 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:54:13 localhost nova_compute[284026]: 2025-11-27 09:54:13.752 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:54:14 localhost ceph-mon[291772]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:54:14 localhost ceph-mon[291772]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:54:14 localhost ceph-mon[291772]: Deploying daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:54:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:14 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:14 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:54:14 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/159594467' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.220 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.468s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:54:14 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.294 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.294 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.496 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.497 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11518MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.498 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.498 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.622 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.622 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.623 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:54:14 localhost nova_compute[284026]: 2025-11-27 09:54:14.681 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:54:15 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:54:15 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1031608707' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:54:15 localhost nova_compute[284026]: 2025-11-27 09:54:15.136 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.455s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:54:15 localhost nova_compute[284026]: 2025-11-27 09:54:15.143 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:54:15 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:54:15 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:54:15 localhost nova_compute[284026]: 2025-11-27 09:54:15.158 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:54:15 localhost nova_compute[284026]: 2025-11-27 09:54:15.160 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:54:15 localhost nova_compute[284026]: 2025-11-27 09:54:15.160 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.662s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:54:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:54:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:54:15 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 adding peer [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] to list of hints Nov 27 04:54:15 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 adding peer [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] to list of hints Nov 27 04:54:16 localhost podman[295204]: 2025-11-27 09:54:16.001642093 +0000 UTC m=+0.091386662 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 04:54:16 localhost podman[295204]: 2025-11-27 09:54:16.013737258 +0000 UTC m=+0.103481817 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, config_id=edpm, managed_by=edpm_ansible) Nov 27 04:54:16 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:54:16 localhost systemd[1]: tmp-crun.bjOhOl.mount: Deactivated successfully. Nov 27 04:54:16 localhost podman[295205]: 2025-11-27 09:54:16.106833465 +0000 UTC m=+0.188480657 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.buildah.version=1.33.7, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., maintainer=Red Hat, Inc., version=9.6, build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, vcs-type=git, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=) Nov 27 04:54:16 localhost podman[295205]: 2025-11-27 09:54:16.123962455 +0000 UTC m=+0.205609647 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, maintainer=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, architecture=x86_64, managed_by=edpm_ansible, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, version=9.6, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:54:16 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:54:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:16 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.162 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.162 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.163 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:54:16 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e8 adding peer [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] to list of hints Nov 27 04:54:16 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x556da9af6000 mon_map magic: 0 from mon.2 v2:172.18.0.108:3300/0 Nov 27 04:54:16 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:54:16 localhost ceph-mon[291772]: paxos.2).electionLogic(44) init, last seen epoch 44 Nov 27 04:54:16 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:16 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.516 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.801 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.802 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.802 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:54:16 localhost nova_compute[284026]: 2025-11-27 09:54:16.803 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:54:17 localhost nova_compute[284026]: 2025-11-27 09:54:17.237 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:54:17 localhost nova_compute[284026]: 2025-11-27 09:54:17.255 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:54:17 localhost nova_compute[284026]: 2025-11-27 09:54:17.256 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:54:18 localhost nova_compute[284026]: 2025-11-27 09:54:18.766 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:54:20 localhost podman[295242]: 2025-11-27 09:54:20.989300168 +0000 UTC m=+0.083981334 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:54:21 localhost podman[295242]: 2025-11-27 09:54:21.020489244 +0000 UTC m=+0.115170370 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:54:21 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:21 localhost nova_compute[284026]: 2025-11-27 09:54:21.547 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:54:21 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:54:21 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537442,np0005537446,np0005537445 in quorum (ranks 0,1,2,3) Nov 27 04:54:21 localhost ceph-mon[291772]: Health check failed: 1/5 mons down, quorum np0005537443,np0005537442,np0005537446,np0005537445 (MON_DOWN) Nov 27 04:54:21 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 1/5 mons down, quorum np0005537443,np0005537442,np0005537446,np0005537445 Nov 27 04:54:21 localhost ceph-mon[291772]: [WRN] MON_DOWN: 1/5 mons down, quorum np0005537443,np0005537442,np0005537446,np0005537445 Nov 27 04:54:21 localhost ceph-mon[291772]: mon.np0005537444 (rank 4) addr [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] is down (out of quorum) Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:22 localhost ceph-mon[291772]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:22 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:22 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:54:23 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: paxos.2).electionLogic(46) init, last seen epoch 46 Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537446@2(electing) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e9 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:54:23 localhost nova_compute[284026]: 2025-11-27 09:54:23.766 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537442 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:54:23 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537442,np0005537446,np0005537445,np0005537444 in quorum (ranks 0,1,2,3,4) Nov 27 04:54:23 localhost ceph-mon[291772]: Health check cleared: MON_DOWN (was: 1/5 mons down, quorum np0005537443,np0005537442,np0005537446,np0005537445) Nov 27 04:54:23 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:54:23 localhost ceph-mon[291772]: overall HEALTH_OK Nov 27 04:54:23 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:23 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:23 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:54:24 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:24 localhost ceph-mon[291772]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:54:24 localhost ceph-mon[291772]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:54:24 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:24 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:24 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:25 localhost openstack_network_exporter[244641]: ERROR 09:54:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:54:25 localhost openstack_network_exporter[244641]: ERROR 09:54:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:54:25 localhost openstack_network_exporter[244641]: ERROR 09:54:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:54:25 localhost openstack_network_exporter[244641]: ERROR 09:54:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:54:25 localhost openstack_network_exporter[244641]: Nov 27 04:54:25 localhost openstack_network_exporter[244641]: ERROR 09:54:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:54:25 localhost openstack_network_exporter[244641]: Nov 27 04:54:25 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:54:25 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:54:25 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:25 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:25 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:26 localhost nova_compute[284026]: 2025-11-27 09:54:26.550 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:26 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:54:26 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:54:26 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:26 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:26 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:27 localhost podman[295320]: Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.18092246 +0000 UTC m=+0.084065976 container create d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, com.redhat.component=rhceph-container, GIT_BRANCH=main, distribution-scope=public, RELEASE=main, release=553, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, version=7, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc.) Nov 27 04:54:27 localhost systemd[1]: Started libpod-conmon-d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94.scope. Nov 27 04:54:27 localhost systemd[1]: Started libcrun container. Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.146497397 +0000 UTC m=+0.049640933 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.257308129 +0000 UTC m=+0.160451645 container init d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, name=rhceph, io.buildah.version=1.33.12, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, version=7, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, architecture=x86_64, GIT_BRANCH=main, GIT_CLEAN=True) Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.2685118 +0000 UTC m=+0.171655316 container start d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, GIT_BRANCH=main, distribution-scope=public, architecture=x86_64, io.openshift.expose-services=, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, vcs-type=git, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, version=7, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, release=553, GIT_CLEAN=True, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.268798257 +0000 UTC m=+0.171941773 container attach d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, name=rhceph, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , architecture=x86_64, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, RELEASE=main) Nov 27 04:54:27 localhost quizzical_gould[295335]: 167 167 Nov 27 04:54:27 localhost systemd[1]: libpod-d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94.scope: Deactivated successfully. Nov 27 04:54:27 localhost podman[295320]: 2025-11-27 09:54:27.273776721 +0000 UTC m=+0.176920237 container died d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, io.openshift.expose-services=, distribution-scope=public, RELEASE=main, ceph=True, architecture=x86_64) Nov 27 04:54:27 localhost podman[295341]: 2025-11-27 09:54:27.371541084 +0000 UTC m=+0.089418180 container remove d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=quizzical_gould, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., GIT_BRANCH=main, RELEASE=main, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, vcs-type=git, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, ceph=True, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, architecture=x86_64, description=Red Hat Ceph Storage 7) Nov 27 04:54:27 localhost systemd[1]: libpod-conmon-d508817dac73a4062c2faa1f315261ed5ad337e0d75a147e8d17c7b1b7c5dd94.scope: Deactivated successfully. Nov 27 04:54:27 localhost ceph-mon[291772]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:54:27 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:54:27 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:27 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:27 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:54:28 localhost podman[295413]: Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.10238411 +0000 UTC m=+0.077962372 container create 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, io.openshift.expose-services=, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, architecture=x86_64, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:54:28 localhost systemd[1]: Started libpod-conmon-3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53.scope. Nov 27 04:54:28 localhost systemd[1]: Started libcrun container. Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.1627559 +0000 UTC m=+0.138334192 container init 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.buildah.version=1.33.12, ceph=True, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, architecture=x86_64, io.openshift.expose-services=, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, release=553, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.0699786 +0000 UTC m=+0.045556922 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.173453307 +0000 UTC m=+0.149031589 container start 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, RELEASE=main, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, release=553, io.buildah.version=1.33.12, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, version=7, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, vcs-type=git, GIT_BRANCH=main, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, name=rhceph, distribution-scope=public) Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.173992072 +0000 UTC m=+0.149570384 container attach 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, GIT_BRANCH=main, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, architecture=x86_64, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, ceph=True, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, distribution-scope=public, io.openshift.tags=rhceph ceph, vcs-type=git) Nov 27 04:54:28 localhost gifted_morse[295429]: 167 167 Nov 27 04:54:28 localhost systemd[1]: libpod-3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53.scope: Deactivated successfully. Nov 27 04:54:28 localhost podman[295413]: 2025-11-27 09:54:28.176483908 +0000 UTC m=+0.152062250 container died 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, CEPH_POINT_RELEASE=, vcs-type=git, RELEASE=main, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, GIT_BRANCH=main, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, GIT_CLEAN=True, release=553, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, version=7, distribution-scope=public, ceph=True, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:54:28 localhost systemd[1]: var-lib-containers-storage-overlay-866fea46f9f101f510ac1ba62199607b0c83ca667ee3179c13635e7905f1eb02-merged.mount: Deactivated successfully. Nov 27 04:54:28 localhost systemd[1]: var-lib-containers-storage-overlay-b500a7574a6af9fb00fc514634de4c2c48e2e32ac176251ba12460dc9e82e1cb-merged.mount: Deactivated successfully. Nov 27 04:54:28 localhost podman[295434]: 2025-11-27 09:54:28.275825973 +0000 UTC m=+0.091785193 container remove 3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_morse, ceph=True, distribution-scope=public, io.buildah.version=1.33.12, vcs-type=git, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, description=Red Hat Ceph Storage 7, RELEASE=main, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, name=rhceph, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., io.openshift.expose-services=, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:54:28 localhost systemd[1]: libpod-conmon-3116ecf93adcd76256c2b8b0446c663c0da9ffc6fe16f7d13fd719aaad62dd53.scope: Deactivated successfully. Nov 27 04:54:28 localhost nova_compute[284026]: 2025-11-27 09:54:28.792 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:28 localhost ceph-mon[291772]: Reconfiguring osd.0 (monmap changed)... Nov 27 04:54:28 localhost ceph-mon[291772]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:28 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost podman[295507]: Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.13899995 +0000 UTC m=+0.075898997 container create 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, vendor=Red Hat, Inc., name=rhceph, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, ceph=True, io.openshift.tags=rhceph ceph, vcs-type=git, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , RELEASE=main, GIT_BRANCH=main) Nov 27 04:54:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:54:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:54:29 localhost systemd[1]: Started libpod-conmon-9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8.scope. Nov 27 04:54:29 localhost systemd[1]: Started libcrun container. Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.108348047 +0000 UTC m=+0.045247084 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.215251715 +0000 UTC m=+0.152150722 container init 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, architecture=x86_64, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, GIT_CLEAN=True, io.buildah.version=1.33.12, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, distribution-scope=public, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, vcs-type=git) Nov 27 04:54:29 localhost systemd[1]: tmp-crun.X8h7XO.mount: Deactivated successfully. Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.230200847 +0000 UTC m=+0.167099884 container start 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., release=553, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_BRANCH=main, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, version=7, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, description=Red Hat Ceph Storage 7) Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.230571396 +0000 UTC m=+0.167470423 container attach 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, distribution-scope=public, name=rhceph, vcs-type=git, release=553, CEPH_POINT_RELEASE=, version=7, GIT_BRANCH=main, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, architecture=x86_64, RELEASE=main, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc.) Nov 27 04:54:29 localhost affectionate_roentgen[295524]: 167 167 Nov 27 04:54:29 localhost systemd[1]: libpod-9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8.scope: Deactivated successfully. Nov 27 04:54:29 localhost podman[295507]: 2025-11-27 09:54:29.233949427 +0000 UTC m=+0.170848514 container died 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, ceph=True, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, name=rhceph, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, architecture=x86_64, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553) Nov 27 04:54:29 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e83 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:29 localhost systemd[1]: tmp-crun.ivzveT.mount: Deactivated successfully. Nov 27 04:54:29 localhost podman[295522]: 2025-11-27 09:54:29.326212382 +0000 UTC m=+0.146124431 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible) Nov 27 04:54:29 localhost podman[295523]: 2025-11-27 09:54:29.295634542 +0000 UTC m=+0.114775290 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:54:29 localhost podman[295522]: 2025-11-27 09:54:29.37310706 +0000 UTC m=+0.193019109 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3) Nov 27 04:54:29 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:54:29 localhost podman[295523]: 2025-11-27 09:54:29.426989956 +0000 UTC m=+0.246130664 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:54:29 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:54:29 localhost podman[295548]: 2025-11-27 09:54:29.493289475 +0000 UTC m=+0.246262618 container remove 9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=affectionate_roentgen, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, RELEASE=main, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, vendor=Red Hat, Inc., architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, ceph=True, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main) Nov 27 04:54:29 localhost systemd[1]: libpod-conmon-9316c710ac55ff125f30a3a5ffaa5de6aad7e38cfdaadd3837fbac88a8ef69c8.scope: Deactivated successfully. Nov 27 04:54:29 localhost ceph-mon[291772]: Reconfiguring osd.3 (monmap changed)... Nov 27 04:54:29 localhost ceph-mon[291772]: Reconfiguring daemon osd.3 on np0005537446.localdomain Nov 27 04:54:29 localhost ceph-mon[291772]: Reconfig service osd.default_drive_group Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' Nov 27 04:54:29 localhost ceph-mon[291772]: from='mgr.14184 172.18.0.105:0/2853074178' entity='mgr.np0005537443.fgpbcx' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:30 localhost systemd[1]: var-lib-containers-storage-overlay-19f85ce0de5936bf94fadca9d9a850f4723ae31a6b1882396a31b79cb6fca018-merged.mount: Deactivated successfully. Nov 27 04:54:30 localhost podman[295649]: Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.374140215 +0000 UTC m=+0.093865450 container create 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, GIT_BRANCH=main, RELEASE=main, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, ceph=True, GIT_CLEAN=True, com.redhat.component=rhceph-container) Nov 27 04:54:30 localhost systemd[1]: Started libpod-conmon-8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e.scope. Nov 27 04:54:30 localhost systemd[1]: Started libcrun container. Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.339420294 +0000 UTC m=+0.059145579 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.443976058 +0000 UTC m=+0.163701293 container init 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., architecture=x86_64, vcs-type=git, ceph=True, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, version=7, CEPH_POINT_RELEASE=, distribution-scope=public, maintainer=Guillaume Abrioux , release=553) Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.454250354 +0000 UTC m=+0.173975599 container start 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, maintainer=Guillaume Abrioux , version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, release=553, name=rhceph, ceph=True, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, vcs-type=git, io.buildah.version=1.33.12, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, distribution-scope=public) Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.454902342 +0000 UTC m=+0.174627627 container attach 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, version=7, architecture=x86_64, release=553, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, RELEASE=main, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, name=rhceph, ceph=True, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc.) Nov 27 04:54:30 localhost gifted_dhawan[295664]: 167 167 Nov 27 04:54:30 localhost systemd[1]: libpod-8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e.scope: Deactivated successfully. Nov 27 04:54:30 localhost podman[295649]: 2025-11-27 09:54:30.460546833 +0000 UTC m=+0.180272088 container died 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, architecture=x86_64, release=553, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , name=rhceph, RELEASE=main, ceph=True, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, GIT_BRANCH=main, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git) Nov 27 04:54:30 localhost podman[295670]: 2025-11-27 09:54:30.559761124 +0000 UTC m=+0.090292113 container remove 8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=gifted_dhawan, release=553, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, maintainer=Guillaume Abrioux , ceph=True, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, vendor=Red Hat, Inc.) Nov 27 04:54:30 localhost systemd[1]: libpod-conmon-8078ae305b183e9c20448447edf974a4887fd0af75fa0f7edb352c1a3ba4df1e.scope: Deactivated successfully. Nov 27 04:54:30 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 e84: 6 total, 6 up, 6 in Nov 27 04:54:30 localhost systemd[1]: session-65.scope: Deactivated successfully. Nov 27 04:54:30 localhost systemd[1]: session-65.scope: Consumed 17.416s CPU time. Nov 27 04:54:30 localhost systemd-logind[761]: Session 65 logged out. Waiting for processes to exit. Nov 27 04:54:30 localhost systemd-logind[761]: Removed session 65. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #16. Immutable memtables: 0. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.720785) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 5] Flushing memtable with next log file: 16 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270720832, "job": 5, "event": "flush_started", "num_memtables": 1, "num_entries": 2914, "num_deletes": 255, "total_data_size": 6031229, "memory_usage": 6106512, "flush_reason": "Manual Compaction"} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 5] Level-0 flush table #17: started Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270750073, "cf_name": "default", "job": 5, "event": "table_file_creation", "file_number": 17, "file_size": 3411415, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 10906, "largest_seqno": 13814, "table_properties": {"data_size": 3399562, "index_size": 7334, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3461, "raw_key_size": 31398, "raw_average_key_size": 22, "raw_value_size": 3373233, "raw_average_value_size": 2465, "num_data_blocks": 319, "num_entries": 1368, "num_filter_entries": 1368, "num_deletions": 254, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237199, "oldest_key_time": 1764237199, "file_creation_time": 1764237270, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 17, "seqno_to_time_mapping": "N/A"}} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 5] Flush lasted 29395 microseconds, and 9815 cpu microseconds. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.750163) [db/flush_job.cc:967] [default] [JOB 5] Level-0 flush table #17: 3411415 bytes OK Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.750205) [db/memtable_list.cc:519] [default] Level-0 commit table #17 started Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.752628) [db/memtable_list.cc:722] [default] Level-0 commit table #17: memtable #1 done Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.752656) EVENT_LOG_v1 {"time_micros": 1764237270752645, "job": 5, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.752689) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 5] Try to delete WAL files size 6017164, prev total WAL file size 6017164, number of live WAL files 2. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000013.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.755693) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003130323931' seq:72057594037927935, type:22 .. '7061786F73003130353433' seq:0, type:0; will stop at (end) Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 6] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 5 Base level 0, inputs: [17(3331KB)], [15(10MB)] Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270755775, "job": 6, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [17], "files_L6": [15], "score": -1, "input_data_size": 13966957, "oldest_snapshot_seqno": -1} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 6] Generated table #18: 9897 keys, 12742375 bytes, temperature: kUnknown Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270848854, "cf_name": "default", "job": 6, "event": "table_file_creation", "file_number": 18, "file_size": 12742375, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 12685410, "index_size": 31136, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 24773, "raw_key_size": 263591, "raw_average_key_size": 26, "raw_value_size": 12515638, "raw_average_value_size": 1264, "num_data_blocks": 1192, "num_entries": 9897, "num_filter_entries": 9897, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237270, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.849165) [db/compaction/compaction_job.cc:1663] [default] [JOB 6] Compacted 1@0 + 1@6 files to L6 => 12742375 bytes Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.857912) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 149.9 rd, 136.8 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(3.3, 10.1 +0.0 blob) out(12.2 +0.0 blob), read-write-amplify(7.8) write-amplify(3.7) OK, records in: 10449, records dropped: 552 output_compression: NoCompression Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.857946) EVENT_LOG_v1 {"time_micros": 1764237270857931, "job": 6, "event": "compaction_finished", "compaction_time_micros": 93170, "compaction_time_cpu_micros": 39508, "output_level": 6, "num_output_files": 1, "total_output_size": 12742375, "num_input_records": 10449, "num_output_records": 9897, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000017.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270858554, "job": 6, "event": "table_file_deletion", "file_number": 17} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000015.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237270860546, "job": 6, "event": "table_file_deletion", "file_number": 15} Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.755511) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.860636) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.860646) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.860650) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.860653) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:54:30.860655) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:54:30 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537446.hkzfou (monmap changed)... Nov 27 04:54:30 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537446.hkzfou on np0005537446.localdomain Nov 27 04:54:30 localhost ceph-mon[291772]: from='client.? 172.18.0.200:0/3544167733' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:54:30 localhost ceph-mon[291772]: Activating manager daemon np0005537441.xbypnw Nov 27 04:54:30 localhost ceph-mon[291772]: from='client.? 172.18.0.200:0/3544167733' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:54:30 localhost ceph-mon[291772]: Manager daemon np0005537441.xbypnw is now available Nov 27 04:54:30 localhost ceph-mon[291772]: removing stray HostCache host record np0005537441.localdomain.devices.0 Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"} : dispatch Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"} : dispatch Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"}]': finished Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"} : dispatch Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"} : dispatch Nov 27 04:54:30 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537441.localdomain.devices.0"}]': finished Nov 27 04:54:31 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537441.xbypnw/mirror_snapshot_schedule"} : dispatch Nov 27 04:54:31 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537441.xbypnw/mirror_snapshot_schedule"} : dispatch Nov 27 04:54:31 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537441.xbypnw/trash_purge_schedule"} : dispatch Nov 27 04:54:31 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537441.xbypnw/trash_purge_schedule"} : dispatch Nov 27 04:54:31 localhost sshd[295687]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:54:31 localhost systemd-logind[761]: New session 66 of user ceph-admin. Nov 27 04:54:31 localhost systemd[1]: Started Session 66 of User ceph-admin. Nov 27 04:54:31 localhost systemd[1]: var-lib-containers-storage-overlay-f71bbd6c103c031ced620686a99b0c2429829c82602eecf6b1aca0d2a2a52dc5-merged.mount: Deactivated successfully. Nov 27 04:54:31 localhost nova_compute[284026]: 2025-11-27 09:54:31.592 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:32 localhost systemd[1]: tmp-crun.5hhwQd.mount: Deactivated successfully. Nov 27 04:54:32 localhost podman[295797]: 2025-11-27 09:54:32.217514557 +0000 UTC m=+0.101557235 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, release=553, name=rhceph, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=) Nov 27 04:54:32 localhost podman[295797]: 2025-11-27 09:54:32.332576244 +0000 UTC m=+0.216618942 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, architecture=x86_64, ceph=True, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., version=7, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, description=Red Hat Ceph Storage 7) Nov 27 04:54:32 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:33 localhost nova_compute[284026]: 2025-11-27 09:54:33.834 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:34 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:35 localhost ceph-mon[291772]: [27/Nov/2025:09:54:33] ENGINE Bus STARTING Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:54:35 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:54:35 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:54:35 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:54:35 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:54:35 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:35 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:35 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:35 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:35 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:54:36 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:36 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:36 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:36 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:36 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:54:36 localhost nova_compute[284026]: 2025-11-27 09:54:36.596 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:54:37 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:37 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537442.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:38 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537442.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:38 localhost podman[242678]: time="2025-11-27T09:54:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:54:38 localhost podman[242678]: @ - - [27/Nov/2025:09:54:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:54:38 localhost podman[242678]: @ - - [27/Nov/2025:09:54:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19694 "" "Go-http-client/1.1" Nov 27 04:54:38 localhost nova_compute[284026]: 2025-11-27 09:54:38.877 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:39 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:39 localhost ceph-mon[291772]: Reconfiguring crash.np0005537442 (monmap changed)... Nov 27 04:54:39 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537442 on np0005537442.localdomain Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537442.byetac", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537442.byetac", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:39 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:40 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537442.byetac (monmap changed)... Nov 27 04:54:40 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537442.byetac on np0005537442.localdomain Nov 27 04:54:40 localhost ceph-mon[291772]: [27/Nov/2025:09:54:39] ENGINE Error in 'start' listener >#012Traceback (most recent call last):#012 File "/lib/python3.9/site-packages/cherrypy/process/wspbus.py", line 230, in publish#012 output.append(listener(*args, **kwargs))#012 File "/lib/python3.9/site-packages/cherrypy/_cpserver.py", line 180, in start#012 super(Server, self).start()#012 File "/lib/python3.9/site-packages/cherrypy/process/servers.py", line 184, in start#012 self.wait()#012 File "/lib/python3.9/site-packages/cherrypy/process/servers.py", line 260, in wait#012 portend.occupied(*self.bound_addr, timeout=Timeouts.occupied)#012 File "/lib/python3.9/site-packages/portend.py", line 162, in occupied#012 raise Timeout("Port {port} not bound on {host}.".format(**locals()))#012portend.Timeout: Port 8765 not bound on 172.18.0.103. Nov 27 04:54:40 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537443.fgpbcx (monmap changed)... Nov 27 04:54:40 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537443.fgpbcx on np0005537443.localdomain Nov 27 04:54:40 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:40 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:40 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:40 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:41 localhost nova_compute[284026]: 2025-11-27 09:54:41.641 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:41 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:54:41 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:54:41 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:41 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:41 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:41 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:54:41 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:41 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:54:41 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:54:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:54:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:54:42 localhost podman[296700]: 2025-11-27 09:54:42.995887529 +0000 UTC m=+0.088605988 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:54:43 localhost podman[296700]: 2025-11-27 09:54:43.010895221 +0000 UTC m=+0.103613710 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:54:43 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:54:43 localhost podman[296699]: 2025-11-27 09:54:43.093697512 +0000 UTC m=+0.187859371 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3) Nov 27 04:54:43 localhost podman[296699]: 2025-11-27 09:54:43.103947267 +0000 UTC m=+0.198109176 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:54:43 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:54:43 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:43 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:43 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:54:43 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:54:43 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:54:43 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:54:43.563 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:54:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:54:43.564 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:54:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:54:43.565 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:54:43 localhost nova_compute[284026]: 2025-11-27 09:54:43.920 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:44 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:44 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:44 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:44 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:44 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:54:44 localhost ceph-mon[291772]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:54:44 localhost ceph-mon[291772]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:54:44 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:45 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:45 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Error in 'start' listener >#012Traceback (most recent call last):#012 File "/lib/python3.9/site-packages/cherrypy/process/wspbus.py", line 230, in publish#012 output.append(listener(*args, **kwargs))#012 File "/lib/python3.9/site-packages/cherrypy/_cpserver.py", line 180, in start#012 super(Server, self).start()#012 File "/lib/python3.9/site-packages/cherrypy/process/servers.py", line 184, in start#012 self.wait()#012 File "/lib/python3.9/site-packages/cherrypy/process/servers.py", line 260, in wait#012 portend.occupied(*self.bound_addr, timeout=Timeouts.occupied)#012 File "/lib/python3.9/site-packages/portend.py", line 162, in occupied#012 raise Timeout("Port {port} not bound on {host}.".format(**locals()))#012portend.Timeout: Port 7150 not bound on 172.18.0.103. Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Shutting down due to error in start listener:#012Traceback (most recent call last):#012 File "/lib/python3.9/site-packages/cherrypy/process/wspbus.py", line 268, in start#012 self.publish('start')#012 File "/lib/python3.9/site-packages/cherrypy/process/wspbus.py", line 248, in publish#012 raise exc#012cherrypy.process.wspbus.ChannelFailures: Timeout('Port 8765 not bound on 172.18.0.103.')#012Timeout('Port 7150 not bound on 172.18.0.103.') Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Bus STOPPING Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('172.18.0.103', 8765)) already shut down Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('172.18.0.103', 7150)) already shut down Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Bus STOPPED Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Bus EXITING Nov 27 04:54:45 localhost ceph-mon[291772]: [27/Nov/2025:09:54:44] ENGINE Bus EXITED Nov 27 04:54:45 localhost ceph-mon[291772]: Failed to run cephadm http server: Timeout('Port 8765 not bound on 172.18.0.103.')#012Timeout('Port 7150 not bound on 172.18.0.103.') Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:45 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:54:46 localhost ceph-mon[291772]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:54:46 localhost nova_compute[284026]: 2025-11-27 09:54:46.645 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:54:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:54:47 localhost podman[296736]: 2025-11-27 09:54:47.006775058 +0000 UTC m=+0.088678480 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:54:47 localhost podman[296737]: 2025-11-27 09:54:47.056890042 +0000 UTC m=+0.135888786 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, com.redhat.component=ubi9-minimal-container, config_id=edpm, maintainer=Red Hat, Inc., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, architecture=x86_64, name=ubi9-minimal) Nov 27 04:54:47 localhost podman[296737]: 2025-11-27 09:54:47.073030875 +0000 UTC m=+0.152029619 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, architecture=x86_64, io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, name=ubi9-minimal, vcs-type=git, release=1755695350, distribution-scope=public, url=https://catalog.redhat.com/en/search?searchType=containers, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, config_id=edpm, managed_by=edpm_ansible, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 04:54:47 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:54:47 localhost podman[296736]: 2025-11-27 09:54:47.124145066 +0000 UTC m=+0.206048528 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:54:47 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:54:47 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:47 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:47 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:47 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:47 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:54:47 localhost ceph-mon[291772]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:54:48 localhost podman[296827]: Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.240157867 +0000 UTC m=+0.082946437 container create 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, build-date=2025-09-24T08:57:55, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, name=rhceph, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, com.redhat.component=rhceph-container, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , release=553, version=7, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:54:48 localhost systemd[1]: Started libpod-conmon-7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6.scope. Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.205668681 +0000 UTC m=+0.048457281 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:48 localhost systemd[1]: Started libcrun container. Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.32345397 +0000 UTC m=+0.166242540 container init 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, version=7, vcs-type=git, GIT_CLEAN=True, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., ceph=True, architecture=x86_64, name=rhceph, io.buildah.version=1.33.12, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux ) Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.33569646 +0000 UTC m=+0.178485030 container start 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, vcs-type=git, RELEASE=main, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , distribution-scope=public, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, CEPH_POINT_RELEASE=, release=553, version=7) Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.336060789 +0000 UTC m=+0.178849359 container attach 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, io.k8s.description=Red Hat Ceph Storage 7, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, architecture=x86_64, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., distribution-scope=public, name=rhceph, build-date=2025-09-24T08:57:55, vcs-type=git, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, RELEASE=main, maintainer=Guillaume Abrioux , ceph=True, version=7, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:54:48 localhost sweet_montalcini[296843]: 167 167 Nov 27 04:54:48 localhost systemd[1]: libpod-7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6.scope: Deactivated successfully. Nov 27 04:54:48 localhost podman[296827]: 2025-11-27 09:54:48.340067656 +0000 UTC m=+0.182856236 container died 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, version=7, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, RELEASE=main, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, GIT_BRANCH=main, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, maintainer=Guillaume Abrioux ) Nov 27 04:54:48 localhost podman[296848]: 2025-11-27 09:54:48.447390096 +0000 UTC m=+0.094039104 container remove 7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sweet_montalcini, io.buildah.version=1.33.12, version=7, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, ceph=True, com.redhat.component=rhceph-container, name=rhceph, maintainer=Guillaume Abrioux , distribution-scope=public, vcs-type=git, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:54:48 localhost systemd[1]: libpod-conmon-7ce39d67d2a8c51151616d36eb784acf2c63b70fd76fd60cf993c00a384190b6.scope: Deactivated successfully. Nov 27 04:54:48 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:48 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:48 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:48 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:48 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:54:48 localhost ceph-mon[291772]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:54:48 localhost nova_compute[284026]: 2025-11-27 09:54:48.956 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:49 localhost systemd[1]: var-lib-containers-storage-overlay-c6a7867282319a6ef3abd310807647bdb44ab7261f267e894658e03e1973e055-merged.mount: Deactivated successfully. Nov 27 04:54:49 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:49 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:54:49 localhost podman[296924]: Nov 27 04:54:49 localhost podman[296924]: 2025-11-27 09:54:49.990216465 +0000 UTC m=+0.073516963 container create 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, architecture=x86_64, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., ceph=True, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, distribution-scope=public, version=7, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , io.openshift.expose-services=, io.buildah.version=1.33.12) Nov 27 04:54:50 localhost systemd[1]: Started libpod-conmon-0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712.scope. Nov 27 04:54:50 localhost systemd[1]: Started libcrun container. Nov 27 04:54:50 localhost podman[296924]: 2025-11-27 09:54:50.046640768 +0000 UTC m=+0.129941276 container init 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, description=Red Hat Ceph Storage 7, name=rhceph, vendor=Red Hat, Inc., distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, RELEASE=main, io.openshift.expose-services=, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, release=553, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, version=7, GIT_CLEAN=True) Nov 27 04:54:50 localhost podman[296924]: 2025-11-27 09:54:50.058576649 +0000 UTC m=+0.141877157 container start 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, RELEASE=main, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, maintainer=Guillaume Abrioux , name=rhceph, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, io.openshift.expose-services=, GIT_CLEAN=True, distribution-scope=public, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container) Nov 27 04:54:50 localhost podman[296924]: 2025-11-27 09:54:50.058944779 +0000 UTC m=+0.142245327 container attach 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, release=553, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, vendor=Red Hat, Inc., version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, distribution-scope=public, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, build-date=2025-09-24T08:57:55, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12) Nov 27 04:54:50 localhost podman[296924]: 2025-11-27 09:54:49.960352134 +0000 UTC m=+0.043652662 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:50 localhost condescending_pascal[296939]: 167 167 Nov 27 04:54:50 localhost systemd[1]: libpod-0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712.scope: Deactivated successfully. Nov 27 04:54:50 localhost podman[296924]: 2025-11-27 09:54:50.062543865 +0000 UTC m=+0.145844393 container died 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , architecture=x86_64, vendor=Red Hat, Inc., vcs-type=git, description=Red Hat Ceph Storage 7, name=rhceph, io.openshift.expose-services=, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, release=553, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:54:50 localhost podman[296944]: 2025-11-27 09:54:50.155022156 +0000 UTC m=+0.085763292 container remove 0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=condescending_pascal, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, distribution-scope=public, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7) Nov 27 04:54:50 localhost systemd[1]: libpod-conmon-0c4ef6183ffe33e418d9fe939f7be714d844ca8b461c1ded03e097e3a1c42712.scope: Deactivated successfully. Nov 27 04:54:50 localhost systemd[1]: tmp-crun.Llofrj.mount: Deactivated successfully. Nov 27 04:54:50 localhost systemd[1]: var-lib-containers-storage-overlay-88e683165b629cd7363a30b5c24675af6d29b13d6bd5298fdfdfea69e45868b5-merged.mount: Deactivated successfully. Nov 27 04:54:50 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537446.hkzfou (monmap changed)... Nov 27 04:54:50 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537446.hkzfou on np0005537446.localdomain Nov 27 04:54:50 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:50 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:50 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:50 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:54:50 localhost podman[297014]: Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.87082858 +0000 UTC m=+0.084669844 container create dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, name=rhceph, RELEASE=main, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, distribution-scope=public, ceph=True, GIT_CLEAN=True, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, release=553, version=7, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:54:50 localhost systemd[1]: Started libpod-conmon-dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d.scope. Nov 27 04:54:50 localhost systemd[1]: Started libcrun container. Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.835312676 +0000 UTC m=+0.049153970 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.938563796 +0000 UTC m=+0.152405070 container init dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, vcs-type=git, GIT_CLEAN=True, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, version=7, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, release=553, name=rhceph, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, ceph=True) Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.947563178 +0000 UTC m=+0.161404452 container start dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, distribution-scope=public, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, GIT_CLEAN=True, io.openshift.expose-services=, RELEASE=main, name=rhceph, architecture=x86_64, vcs-type=git, ceph=True, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7) Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.947770083 +0000 UTC m=+0.161611347 container attach dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, RELEASE=main, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, architecture=x86_64, vcs-type=git, version=7, distribution-scope=public, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., io.buildah.version=1.33.12, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True) Nov 27 04:54:50 localhost sharp_heyrovsky[297029]: 167 167 Nov 27 04:54:50 localhost systemd[1]: libpod-dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d.scope: Deactivated successfully. Nov 27 04:54:50 localhost podman[297014]: 2025-11-27 09:54:50.951431011 +0000 UTC m=+0.165272305 container died dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, version=7, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, io.openshift.expose-services=, name=rhceph, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, release=553, GIT_BRANCH=main, RELEASE=main) Nov 27 04:54:51 localhost podman[297034]: 2025-11-27 09:54:51.050045977 +0000 UTC m=+0.086681786 container remove dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sharp_heyrovsky, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, RELEASE=main, vcs-type=git, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, GIT_CLEAN=True, release=553, GIT_BRANCH=main, distribution-scope=public, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph) Nov 27 04:54:51 localhost systemd[1]: libpod-conmon-dd8296034c3b5ba162d67f1fab53ae8caa9c8bc07ef801176922492277f3c76d.scope: Deactivated successfully. Nov 27 04:54:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:54:51 localhost podman[297050]: 2025-11-27 09:54:51.174386753 +0000 UTC m=+0.083582974 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:54:51 localhost podman[297050]: 2025-11-27 09:54:51.191010139 +0000 UTC m=+0.100206420 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:54:51 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:54:51 localhost systemd[1]: var-lib-containers-storage-overlay-0652b4688c04b58afb9190238bf209c7f835e57650f6a401bed480476ad334c9-merged.mount: Deactivated successfully. Nov 27 04:54:51 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537446.nfulyg (monmap changed)... Nov 27 04:54:51 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537446.nfulyg on np0005537446.localdomain Nov 27 04:54:51 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:51 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:51 localhost ceph-mon[291772]: from='mgr.17370 172.18.0.103:0/93985291' entity='mgr.np0005537441.xbypnw' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:54:51 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:51 localhost nova_compute[284026]: 2025-11-27 09:54:51.691 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:52 localhost ceph-mon[291772]: Health check failed: 1 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:54:52 localhost ceph-mon[291772]: Health check failed: 1 stray host(s) with 1 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:54:53 localhost nova_compute[284026]: 2025-11-27 09:54:53.997 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:54 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:55 localhost openstack_network_exporter[244641]: ERROR 09:54:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:54:55 localhost openstack_network_exporter[244641]: ERROR 09:54:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:54:55 localhost openstack_network_exporter[244641]: ERROR 09:54:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:54:55 localhost openstack_network_exporter[244641]: ERROR 09:54:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:54:55 localhost openstack_network_exporter[244641]: Nov 27 04:54:55 localhost openstack_network_exporter[244641]: ERROR 09:54:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:54:55 localhost openstack_network_exporter[244641]: Nov 27 04:54:56 localhost ceph-mon[291772]: from='mgr.17370 ' entity='mgr.np0005537441.xbypnw' Nov 27 04:54:56 localhost nova_compute[284026]: 2025-11-27 09:54:56.694 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:59 localhost nova_compute[284026]: 2025-11-27 09:54:59.034 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:54:59 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:54:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:54:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:55:00 localhost podman[297090]: 2025-11-27 09:55:00.003673547 +0000 UTC m=+0.091865636 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true, config_id=ovn_controller, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_controller) Nov 27 04:55:00 localhost podman[297090]: 2025-11-27 09:55:00.042980521 +0000 UTC m=+0.131172660 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, config_id=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:55:00 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:55:00 localhost podman[297091]: 2025-11-27 09:55:00.097790201 +0000 UTC m=+0.186110863 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:55:00 localhost podman[297091]: 2025-11-27 09:55:00.109900687 +0000 UTC m=+0.198221349 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:55:00 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:55:01 localhost nova_compute[284026]: 2025-11-27 09:55:01.733 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:04 localhost nova_compute[284026]: 2025-11-27 09:55:04.072 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:04 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:06 localhost nova_compute[284026]: 2025-11-27 09:55:06.774 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:08 localhost podman[242678]: time="2025-11-27T09:55:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:55:08 localhost podman[242678]: @ - - [27/Nov/2025:09:55:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:55:08 localhost podman[242678]: @ - - [27/Nov/2025:09:55:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19695 "" "Go-http-client/1.1" Nov 27 04:55:09 localhost nova_compute[284026]: 2025-11-27 09:55:09.115 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:09 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:11 localhost nova_compute[284026]: 2025-11-27 09:55:11.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:11 localhost nova_compute[284026]: 2025-11-27 09:55:11.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:11 localhost nova_compute[284026]: 2025-11-27 09:55:11.806 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:12 localhost nova_compute[284026]: 2025-11-27 09:55:12.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:12 localhost nova_compute[284026]: 2025-11-27 09:55:12.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:12 localhost nova_compute[284026]: 2025-11-27 09:55:12.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:13 localhost nova_compute[284026]: 2025-11-27 09:55:13.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:13 localhost nova_compute[284026]: 2025-11-27 09:55:13.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:13 localhost nova_compute[284026]: 2025-11-27 09:55:13.731 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:55:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:55:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:55:13 localhost podman[297138]: 2025-11-27 09:55:13.995939908 +0000 UTC m=+0.084615091 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.build-date=20251125, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 04:55:14 localhost podman[297137]: 2025-11-27 09:55:14.041363117 +0000 UTC m=+0.133025490 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 04:55:14 localhost podman[297138]: 2025-11-27 09:55:14.058546878 +0000 UTC m=+0.147222081 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, container_name=multipathd) Nov 27 04:55:14 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:55:14 localhost podman[297137]: 2025-11-27 09:55:14.072447561 +0000 UTC m=+0.164109984 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 04:55:14 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:55:14 localhost nova_compute[284026]: 2025-11-27 09:55:14.148 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:14 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e84 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:14 localhost nova_compute[284026]: 2025-11-27 09:55:14.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:14 localhost nova_compute[284026]: 2025-11-27 09:55:14.767 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:14 localhost nova_compute[284026]: 2025-11-27 09:55:14.767 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:55:14 localhost nova_compute[284026]: 2025-11-27 09:55:14.768 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:55:15 localhost nova_compute[284026]: 2025-11-27 09:55:15.827 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:55:15 localhost nova_compute[284026]: 2025-11-27 09:55:15.827 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:55:15 localhost nova_compute[284026]: 2025-11-27 09:55:15.828 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:55:15 localhost nova_compute[284026]: 2025-11-27 09:55:15.828 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:55:16 localhost nova_compute[284026]: 2025-11-27 09:55:16.847 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.532 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.551 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.551 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.552 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.568 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.569 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.569 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.570 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:55:17 localhost nova_compute[284026]: 2025-11-27 09:55:17.570 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:55:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:55:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:55:17 localhost systemd[1]: tmp-crun.AT2XxN.mount: Deactivated successfully. Nov 27 04:55:17 localhost podman[297197]: 2025-11-27 09:55:17.996639676 +0000 UTC m=+0.087610282 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, url=https://catalog.redhat.com/en/search?searchType=containers, maintainer=Red Hat, Inc., release=1755695350, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, version=9.6, container_name=openstack_network_exporter, io.openshift.expose-services=, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, managed_by=edpm_ansible, architecture=x86_64, com.redhat.component=ubi9-minimal-container, build-date=2025-08-20T13:12:41, distribution-scope=public, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 04:55:18 localhost podman[297197]: 2025-11-27 09:55:18.010923809 +0000 UTC m=+0.101894455 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, name=ubi9-minimal, release=1755695350, architecture=x86_64, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, maintainer=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, config_id=edpm, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.expose-services=, distribution-scope=public, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:55:18 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:55:18 localhost podman[297196]: 2025-11-27 09:55:17.976513575 +0000 UTC m=+0.073430341 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_id=edpm, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:55:18 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e9 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:55:18 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/152734569' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.068 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.498s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:55:18 localhost podman[297196]: 2025-11-27 09:55:18.072706006 +0000 UTC m=+0.169622702 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_managed=true, config_id=edpm, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 04:55:18 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.120 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.120 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.275 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.275 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11520MB free_disk=0.0GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.276 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.276 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.345 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.345 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.345 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=0GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.380 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.841 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.460s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.848 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'MEMORY_MB': {'total': 15738, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 512}, 'VCPU': {'total': 8, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0, 'reserved': 0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:55:18 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e85 e85: 6 total, 6 up, 6 in Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.892 284030 ERROR nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [req-7a5c5220-4eed-4007-8e42-b3e14915436e] Failed to update inventory to [{'MEMORY_MB': {'total': 15738, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0, 'reserved': 512}, 'VCPU': {'total': 8, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0, 'reserved': 0}}] for resource provider with UUID 5764ad4c-cf89-4d5f-a185-92be71f7b67f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n update conflict: Inventory for 'DISK_GB' on resource provider '5764ad4c-cf89-4d5f-a185-92be71f7b67f' in use. ", "code": "placement.inventory.inuse", "request_id": "req-7a5c5220-4eed-4007-8e42-b3e14915436e"}]}#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.899 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.623s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Error updating PCI resources for node np0005537446.localdomain.: nova.exception.PlacementPciException: Failed to gather or report PCI resources to Placement: There was a conflict when trying to complete your request. Nov 27 04:55:18 localhost nova_compute[284026]: Nov 27 04:55:18 localhost nova_compute[284026]: update conflict: Inventory for 'DISK_GB' on resource provider '5764ad4c-cf89-4d5f-a185-92be71f7b67f' in use. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Traceback (most recent call last): Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py", line 1288, in _update_to_placement Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self.reportclient.update_from_provider_tree( Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/scheduler/client/report.py", line 1484, in update_from_provider_tree Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self.set_inventory_for_provider( Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/scheduler/client/report.py", line 987, in set_inventory_for_provider Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager raise exception.InventoryInUse(err['detail']) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager nova.exception.InventoryInUse: There was a conflict when trying to complete your request. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager update conflict: Inventory for 'DISK_GB' on resource provider '5764ad4c-cf89-4d5f-a185-92be71f7b67f' in use. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager During handling of the above exception, another exception occurred: Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Traceback (most recent call last): Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/manager.py", line 10513, in _update_available_resource_for_node Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py", line 889, in update_available_resource Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py", line 414, in inner Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager return f(*args, **kwargs) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py", line 994, in _update_available_resource Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self._update(context, cn, startup=startup) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py", line 1303, in _update Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/retrying.py", line 49, in wrapped_f Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/retrying.py", line 206, in call Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager return attempt.get(self._wrap_exception) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/retrying.py", line 247, in get Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager six.reraise(self.value[0], self.value[1], self.value[2]) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/six.py", line 709, in reraise Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager raise value Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/retrying.py", line 200, in call Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager File "/usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py", line 1298, in _update_to_placement Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager raise exception.PlacementPciException(error=str(e)) Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager nova.exception.PlacementPciException: Failed to gather or report PCI resources to Placement: There was a conflict when trying to complete your request. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager update conflict: Inventory for 'DISK_GB' on resource provider '5764ad4c-cf89-4d5f-a185-92be71f7b67f' in use. Nov 27 04:55:18 localhost nova_compute[284026]: 2025-11-27 09:55:18.900 284030 ERROR nova.compute.manager #033[00m Nov 27 04:55:18 localhost systemd[1]: session-66.scope: Deactivated successfully. Nov 27 04:55:18 localhost systemd[1]: session-66.scope: Consumed 8.330s CPU time. Nov 27 04:55:18 localhost systemd-logind[761]: Session 66 logged out. Waiting for processes to exit. Nov 27 04:55:18 localhost systemd-logind[761]: Removed session 66. Nov 27 04:55:19 localhost nova_compute[284026]: 2025-11-27 09:55:19.153 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:19 localhost sshd[297259]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:55:19 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e85 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:19 localhost systemd-logind[761]: New session 67 of user ceph-admin. Nov 27 04:55:19 localhost systemd[1]: Started Session 67 of User ceph-admin. Nov 27 04:55:19 localhost ceph-mon[291772]: from='client.? 172.18.0.200:0/290462773' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:55:19 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:55:19 localhost ceph-mon[291772]: Activating manager daemon np0005537442.byetac Nov 27 04:55:19 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:55:19 localhost ceph-mon[291772]: Manager daemon np0005537442.byetac is now available Nov 27 04:55:19 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537442.byetac/mirror_snapshot_schedule"} : dispatch Nov 27 04:55:19 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537442.byetac/mirror_snapshot_schedule"} : dispatch Nov 27 04:55:19 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537442.byetac/trash_purge_schedule"} : dispatch Nov 27 04:55:19 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537442.byetac/trash_purge_schedule"} : dispatch Nov 27 04:55:20 localhost systemd[1]: tmp-crun.d6f99p.mount: Deactivated successfully. Nov 27 04:55:20 localhost podman[297370]: 2025-11-27 09:55:20.40185118 +0000 UTC m=+0.107620328 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, RELEASE=main, ceph=True, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, release=553, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, vcs-type=git, io.buildah.version=1.33.12) Nov 27 04:55:20 localhost podman[297370]: 2025-11-27 09:55:20.52302619 +0000 UTC m=+0.228795338 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., GIT_BRANCH=main, name=rhceph, architecture=x86_64, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, build-date=2025-09-24T08:57:55, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, version=7, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:55:20 localhost ceph-mon[291772]: [27/Nov/2025:09:55:20] ENGINE Bus STARTING Nov 27 04:55:20 localhost ceph-mon[291772]: [27/Nov/2025:09:55:20] ENGINE Serving on https://172.18.0.104:7150 Nov 27 04:55:20 localhost ceph-mon[291772]: [27/Nov/2025:09:55:20] ENGINE Client ('172.18.0.104', 57546) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:55:20 localhost ceph-mon[291772]: [27/Nov/2025:09:55:20] ENGINE Serving on http://172.18.0.104:8765 Nov 27 04:55:20 localhost ceph-mon[291772]: [27/Nov/2025:09:55:20] ENGINE Bus STARTED Nov 27 04:55:20 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:20 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #19. Immutable memtables: 0. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.731803) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 7] Flushing memtable with next log file: 19 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321731899, "job": 7, "event": "flush_started", "num_memtables": 1, "num_entries": 1763, "num_deletes": 258, "total_data_size": 7202312, "memory_usage": 7838304, "flush_reason": "Manual Compaction"} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 7] Level-0 flush table #20: started Nov 27 04:55:21 localhost podman[297510]: 2025-11-27 09:55:21.753186092 +0000 UTC m=+0.090071428 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321767158, "cf_name": "default", "job": 7, "event": "table_file_creation", "file_number": 20, "file_size": 4392096, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 13819, "largest_seqno": 15577, "table_properties": {"data_size": 4384522, "index_size": 4271, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2245, "raw_key_size": 19272, "raw_average_key_size": 21, "raw_value_size": 4368065, "raw_average_value_size": 4941, "num_data_blocks": 181, "num_entries": 884, "num_filter_entries": 884, "num_deletions": 258, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237270, "oldest_key_time": 1764237270, "file_creation_time": 1764237321, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 20, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 7] Flush lasted 35739 microseconds, and 9571 cpu microseconds. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.767550) [db/flush_job.cc:967] [default] [JOB 7] Level-0 flush table #20: 4392096 bytes OK Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.767621) [db/memtable_list.cc:519] [default] Level-0 commit table #20 started Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.769949) [db/memtable_list.cc:722] [default] Level-0 commit table #20: memtable #1 done Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.769980) EVENT_LOG_v1 {"time_micros": 1764237321769971, "job": 7, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.770005) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 7] Try to delete WAL files size 7193529, prev total WAL file size 7211305, number of live WAL files 2. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000016.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.771658) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6C6F676D0033353136' seq:72057594037927935, type:22 .. '6C6F676D0033373639' seq:0, type:0; will stop at (end) Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 8] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 7 Base level 0, inputs: [20(4289KB)], [18(12MB)] Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321771856, "job": 8, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [20], "files_L6": [18], "score": -1, "input_data_size": 17134471, "oldest_snapshot_seqno": -1} Nov 27 04:55:21 localhost podman[297510]: 2025-11-27 09:55:21.795008234 +0000 UTC m=+0.131893520 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:55:21 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:55:21 localhost nova_compute[284026]: 2025-11-27 09:55:21.894 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 8] Generated table #21: 10230 keys, 16889657 bytes, temperature: kUnknown Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321904391, "cf_name": "default", "job": 8, "event": "table_file_creation", "file_number": 21, "file_size": 16889657, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 16828281, "index_size": 34715, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 25605, "raw_key_size": 272622, "raw_average_key_size": 26, "raw_value_size": 16650461, "raw_average_value_size": 1627, "num_data_blocks": 1344, "num_entries": 10230, "num_filter_entries": 10230, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237321, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 21, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.904685) [db/compaction/compaction_job.cc:1663] [default] [JOB 8] Compacted 1@0 + 1@6 files to L6 => 16889657 bytes Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.906255) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 129.2 rd, 127.3 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(4.2, 12.2 +0.0 blob) out(16.1 +0.0 blob), read-write-amplify(7.7) write-amplify(3.8) OK, records in: 10781, records dropped: 551 output_compression: NoCompression Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.906276) EVENT_LOG_v1 {"time_micros": 1764237321906267, "job": 8, "event": "compaction_finished", "compaction_time_micros": 132669, "compaction_time_cpu_micros": 44756, "output_level": 6, "num_output_files": 1, "total_output_size": 16889657, "num_input_records": 10781, "num_output_records": 10230, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000020.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321906810, "job": 8, "event": "table_file_deletion", "file_number": 20} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000018.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321908155, "job": 8, "event": "table_file_deletion", "file_number": 18} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.771487) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908208) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908216) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908219) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908222) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908225) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #22. Immutable memtables: 0. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.908803) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 9] Flushing memtable with next log file: 22 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321908882, "job": 9, "event": "flush_started", "num_memtables": 1, "num_entries": 261, "num_deletes": 250, "total_data_size": 19829, "memory_usage": 26904, "flush_reason": "Manual Compaction"} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 9] Level-0 flush table #23: started Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321912092, "cf_name": "default", "job": 9, "event": "table_file_creation", "file_number": 23, "file_size": 12912, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 15579, "largest_seqno": 15838, "table_properties": {"data_size": 11107, "index_size": 49, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 709, "raw_key_size": 3982, "raw_average_key_size": 15, "raw_value_size": 7686, "raw_average_value_size": 29, "num_data_blocks": 2, "num_entries": 260, "num_filter_entries": 260, "num_deletions": 250, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237321, "oldest_key_time": 1764237321, "file_creation_time": 1764237321, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 9] Flush lasted 3368 microseconds, and 827 cpu microseconds. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.912167) [db/flush_job.cc:967] [default] [JOB 9] Level-0 flush table #23: 12912 bytes OK Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.912215) [db/memtable_list.cc:519] [default] Level-0 commit table #23 started Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.913853) [db/memtable_list.cc:722] [default] Level-0 commit table #23: memtable #1 done Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.913876) EVENT_LOG_v1 {"time_micros": 1764237321913870, "job": 9, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.913895) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 9] Try to delete WAL files size 17776, prev total WAL file size 17776, number of live WAL files 2. Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.914873) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6B760031303232' seq:72057594037927935, type:22 .. '6B760031323733' seq:0, type:0; will stop at (end) Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 10] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 9 Base level 0, inputs: [23(12KB)], [21(16MB)] Nov 27 04:55:21 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237321914904, "job": 10, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [23], "files_L6": [21], "score": -1, "input_data_size": 16902569, "oldest_snapshot_seqno": -1} Nov 27 04:55:21 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_DAEMON (was: 1 stray daemon(s) not managed by cephadm) Nov 27 04:55:21 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_HOST (was: 1 stray host(s) with 1 daemon(s) not managed by cephadm) Nov 27 04:55:21 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:21 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 10] Generated table #24: 9981 keys, 15967985 bytes, temperature: kUnknown Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237322036923, "cf_name": "default", "job": 10, "event": "table_file_creation", "file_number": 24, "file_size": 15967985, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 15909567, "index_size": 32397, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 24965, "raw_key_size": 268891, "raw_average_key_size": 26, "raw_value_size": 15737204, "raw_average_value_size": 1576, "num_data_blocks": 1227, "num_entries": 9981, "num_filter_entries": 9981, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237321, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 24, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.037211) [db/compaction/compaction_job.cc:1663] [default] [JOB 10] Compacted 1@0 + 1@6 files to L6 => 15967985 bytes Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.039036) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 138.4 rd, 130.8 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.0, 16.1 +0.0 blob) out(15.2 +0.0 blob), read-write-amplify(2545.7) write-amplify(1236.7) OK, records in: 10490, records dropped: 509 output_compression: NoCompression Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.039059) EVENT_LOG_v1 {"time_micros": 1764237322039048, "job": 10, "event": "compaction_finished", "compaction_time_micros": 122114, "compaction_time_cpu_micros": 28901, "output_level": 6, "num_output_files": 1, "total_output_size": 15967985, "num_input_records": 10490, "num_output_records": 9981, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000023.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237322039162, "job": 10, "event": "table_file_deletion", "file_number": 23} Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000021.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237322040816, "job": 10, "event": "table_file_deletion", "file_number": 21} Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:21.914797) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.040855) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.040861) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.040865) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.040868) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:22 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:22.040871) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:55:23 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:24 localhost nova_compute[284026]: 2025-11-27 09:55:24.198 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:24 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e85 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:24 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:55:24 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:55:24 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:55:24 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:24 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:24 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:24 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:24 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:24 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:24 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:24 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:25 localhost openstack_network_exporter[244641]: ERROR 09:55:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:55:25 localhost openstack_network_exporter[244641]: ERROR 09:55:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:55:25 localhost openstack_network_exporter[244641]: ERROR 09:55:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:55:25 localhost openstack_network_exporter[244641]: ERROR 09:55:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:55:25 localhost openstack_network_exporter[244641]: Nov 27 04:55:25 localhost openstack_network_exporter[244641]: ERROR 09:55:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:55:25 localhost openstack_network_exporter[244641]: Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:25 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:25 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:25 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:26 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:26 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:26 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:26 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:26 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:55:26 localhost nova_compute[284026]: 2025-11-27 09:55:26.928 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:27 localhost ceph-mon[291772]: Reconfiguring mon.np0005537442 (monmap changed)... Nov 27 04:55:27 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537442 on np0005537442.localdomain Nov 27 04:55:27 localhost ceph-mon[291772]: Health check failed: 1 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:55:27 localhost ceph-mon[291772]: Health check failed: 1 stray host(s) with 1 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:55:27 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:27 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:27 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:55:28 localhost ceph-mon[291772]: Reconfiguring mon.np0005537443 (monmap changed)... Nov 27 04:55:28 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537443 on np0005537443.localdomain Nov 27 04:55:28 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:28 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:28 localhost ceph-mon[291772]: Reconfiguring mon.np0005537444 (monmap changed)... Nov 27 04:55:28 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:55:28 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:55:29 localhost nova_compute[284026]: 2025-11-27 09:55:29.200 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:29 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e85 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:30 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:30 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:30 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:30 localhost ceph-mon[291772]: Reconfiguring mon.np0005537445 (monmap changed)... Nov 27 04:55:30 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:55:30 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:55:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:55:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:55:30 localhost podman[298313]: 2025-11-27 09:55:30.434951469 +0000 UTC m=+0.098667039 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:55:30 localhost podman[298313]: 2025-11-27 09:55:30.450941807 +0000 UTC m=+0.114657437 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:55:30 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:55:30 localhost podman[298312]: 2025-11-27 09:55:30.541565629 +0000 UTC m=+0.205492905 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=ovn_controller, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:55:30 localhost podman[298312]: 2025-11-27 09:55:30.582040694 +0000 UTC m=+0.245967950 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, config_id=ovn_controller, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 04:55:30 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:55:30 localhost podman[298390]: Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.86031407 +0000 UTC m=+0.077351837 container create 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, GIT_BRANCH=main, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.openshift.expose-services=, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, ceph=True, vcs-type=git, name=rhceph, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, com.redhat.component=rhceph-container) Nov 27 04:55:30 localhost systemd[1]: Started libpod-conmon-45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063.scope. Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.829727469 +0000 UTC m=+0.046765266 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:55:30 localhost systemd[1]: Started libcrun container. Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.948595418 +0000 UTC m=+0.165633205 container init 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, architecture=x86_64, release=553, io.buildah.version=1.33.12, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, distribution-scope=public, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, CEPH_POINT_RELEASE=, ceph=True, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.959025668 +0000 UTC m=+0.176063435 container start 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, RELEASE=main, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, GIT_CLEAN=True, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., ceph=True, io.buildah.version=1.33.12, GIT_BRANCH=main, io.openshift.expose-services=) Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.959215673 +0000 UTC m=+0.176253530 container attach 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, GIT_CLEAN=True, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, GIT_BRANCH=main, name=rhceph, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, architecture=x86_64, io.buildah.version=1.33.12) Nov 27 04:55:30 localhost great_mendel[298405]: 167 167 Nov 27 04:55:30 localhost systemd[1]: libpod-45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063.scope: Deactivated successfully. Nov 27 04:55:30 localhost podman[298390]: 2025-11-27 09:55:30.964173486 +0000 UTC m=+0.181211293 container died 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, io.buildah.version=1.33.12, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, name=rhceph, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, distribution-scope=public, release=553, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git) Nov 27 04:55:31 localhost podman[298410]: 2025-11-27 09:55:31.046117874 +0000 UTC m=+0.075225239 container remove 45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=great_mendel, io.buildah.version=1.33.12, architecture=x86_64, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., ceph=True, io.openshift.expose-services=, CEPH_POINT_RELEASE=, vcs-type=git, distribution-scope=public, io.openshift.tags=rhceph ceph, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:55:31 localhost systemd[1]: libpod-conmon-45f5e45e2d4ab5fe29584e2cc3f25ee8084a28033b37d0a3d0c6f76f1b59f063.scope: Deactivated successfully. Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:31 localhost ceph-mon[291772]: Reconfiguring mon.np0005537446 (monmap changed)... Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:55:31 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 172.18.0.104:0/2716180017' entity='mgr.np0005537442.byetac' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:31 localhost ceph-mon[291772]: from='mgr.24104 ' entity='mgr.np0005537442.byetac' Nov 27 04:55:31 localhost systemd[1]: var-lib-containers-storage-overlay-f2b9577b06af7c627cd79a8a41f87bb88e18ce0eb7a35295698b0530f09c104f-merged.mount: Deactivated successfully. Nov 27 04:55:31 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e9 handle_command mon_command({"prefix": "mgr fail"} v 0) Nov 27 04:55:31 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='client.? 172.18.0.200:0/4044200370' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:55:31 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e86 e86: 6 total, 6 up, 6 in Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr handle_mgr_map Activating! Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr handle_mgr_map I am now activating Nov 27 04:55:31 localhost ceph-mgr[290377]: [balancer DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: balancer Nov 27 04:55:31 localhost ceph-mgr[290377]: [balancer INFO root] Starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [cephadm DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_09:55:31 Nov 27 04:55:31 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 04:55:31 localhost ceph-mgr[290377]: [balancer INFO root] Some PGs (1.000000) are unknown; try again later Nov 27 04:55:31 localhost systemd[1]: session-67.scope: Deactivated successfully. Nov 27 04:55:31 localhost systemd[1]: session-67.scope: Consumed 6.812s CPU time. Nov 27 04:55:31 localhost systemd-logind[761]: Session 67 logged out. Waiting for processes to exit. Nov 27 04:55:31 localhost systemd-logind[761]: Removed session 67. Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: cephadm Nov 27 04:55:31 localhost ceph-mgr[290377]: [crash DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: crash Nov 27 04:55:31 localhost ceph-mgr[290377]: [devicehealth DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: devicehealth Nov 27 04:55:31 localhost ceph-mgr[290377]: [devicehealth INFO root] Starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [iostat DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: iostat Nov 27 04:55:31 localhost ceph-mgr[290377]: [nfs DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: nfs Nov 27 04:55:31 localhost ceph-mgr[290377]: [orchestrator DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: orchestrator Nov 27 04:55:31 localhost ceph-mgr[290377]: [pg_autoscaler DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: pg_autoscaler Nov 27 04:55:31 localhost ceph-mgr[290377]: [progress DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: progress Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: [progress INFO root] Loading... Nov 27 04:55:31 localhost ceph-mgr[290377]: [progress INFO root] Loaded [, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ] historic events Nov 27 04:55:31 localhost ceph-mgr[290377]: [progress INFO root] Loaded OSDMap, ready. Nov 27 04:55:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] recovery thread starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] starting setup Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: rbd_support Nov 27 04:55:31 localhost ceph-mgr[290377]: [restful DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: restful Nov 27 04:55:31 localhost ceph-mgr[290377]: [restful INFO root] server_addr: :: server_port: 8003 Nov 27 04:55:31 localhost ceph-mgr[290377]: [status DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: status Nov 27 04:55:31 localhost ceph-mgr[290377]: [restful WARNING root] server not running: no certificate configured Nov 27 04:55:31 localhost ceph-mgr[290377]: [telemetry DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: telemetry Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [volumes DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:55:31 localhost nova_compute[284026]: 2025-11-27 09:55:31.931 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 04:55:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 04:55:31 localhost ceph-mgr[290377]: mgr load Constructed class from module: volumes Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] PerfHandler: starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: vms, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: volumes, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.966+0000 7fadee43d640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.966+0000 7fadee43d640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.966+0000 7fadee43d640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.966+0000 7fadee43d640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.966+0000 7fadee43d640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: images, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.970+0000 7fadf1c44640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.970+0000 7fadf1c44640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.970+0000 7fadf1c44640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.970+0000 7fadf1c44640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:55:31.970+0000 7fadf1c44640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: backups, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] TaskHandler: starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: starting Nov 27 04:55:31 localhost ceph-mgr[290377]: [rbd_support INFO root] setup complete Nov 27 04:55:32 localhost sshd[298587]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:55:32 localhost systemd-logind[761]: New session 68 of user ceph-admin. Nov 27 04:55:32 localhost systemd[1]: Started Session 68 of User ceph-admin. Nov 27 04:55:32 localhost ceph-mon[291772]: from='client.? 172.18.0.200:0/4044200370' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:55:32 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:55:32 localhost ceph-mon[291772]: Activating manager daemon np0005537446.nfulyg Nov 27 04:55:32 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:55:32 localhost ceph-mon[291772]: Manager daemon np0005537446.nfulyg is now available Nov 27 04:55:32 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/mirror_snapshot_schedule"} : dispatch Nov 27 04:55:32 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/trash_purge_schedule"} : dispatch Nov 27 04:55:32 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v3: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:33 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:55:33] ENGINE Bus STARTING Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:55:33] ENGINE Bus STARTING Nov 27 04:55:33 localhost systemd[1]: tmp-crun.RPAdHN.mount: Deactivated successfully. Nov 27 04:55:33 localhost podman[298699]: 2025-11-27 09:55:33.239630279 +0000 UTC m=+0.105632245 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, distribution-scope=public, build-date=2025-09-24T08:57:55, architecture=x86_64, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, GIT_BRANCH=main, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, CEPH_POINT_RELEASE=, RELEASE=main, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:55:33 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:55:33] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:55:33] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:55:33 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:55:33] ENGINE Client ('172.18.0.108', 41148) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:55:33] ENGINE Client ('172.18.0.108', 41148) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:55:33 localhost podman[298699]: 2025-11-27 09:55:33.344147433 +0000 UTC m=+0.210149359 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, release=553, vcs-type=git, com.redhat.component=rhceph-container, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., version=7, RELEASE=main, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main) Nov 27 04:55:33 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:55:33] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:55:33] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:55:33 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:55:33] ENGINE Bus STARTED Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:55:33] ENGINE Bus STARTED Nov 27 04:55:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v4: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:33 localhost ceph-mon[291772]: [27/Nov/2025:09:55:33] ENGINE Bus STARTING Nov 27 04:55:33 localhost ceph-mon[291772]: [27/Nov/2025:09:55:33] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:55:33 localhost ceph-mon[291772]: [27/Nov/2025:09:55:33] ENGINE Client ('172.18.0.108', 41148) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:55:33 localhost ceph-mon[291772]: [27/Nov/2025:09:55:33] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:55:33 localhost ceph-mon[291772]: [27/Nov/2025:09:55:33] ENGINE Bus STARTED Nov 27 04:55:33 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:33 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:33 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:33 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mgr[290377]: [devicehealth INFO root] Check health Nov 27 04:55:34 localhost nova_compute[284026]: 2025-11-27 09:55:34.231 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:34 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:34 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_DAEMON (was: 1 stray daemon(s) not managed by cephadm) Nov 27 04:55:34 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_HOST (was: 1 stray host(s) with 1 daemon(s) not managed by cephadm) Nov 27 04:55:34 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:34 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v5: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd/host:np0005537442", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:55:36 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:36 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mgr.np0005537442.byetac 172.18.0.104:0/522867664; not ready for session (expect reconnect) Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:36 localhost nova_compute[284026]: 2025-11-27 09:55:36.966 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:55:37 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:37 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:55:37 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:55:37 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:37 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:37 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:37 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mgr.np0005537442.byetac 172.18.0.104:0/522867664; not ready for session (expect reconnect) Nov 27 04:55:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v6: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:38 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:38 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:38 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:38 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:38 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:38 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v7: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 37 KiB/s rd, 0 B/s wr, 20 op/s Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 5933fb9d-4be7-44fc-bdfb-11cb9ba5a495 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 5933fb9d-4be7-44fc-bdfb-11cb9ba5a495 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] Completed event 5933fb9d-4be7-44fc-bdfb-11cb9ba5a495 (Updating node-proxy deployment (+5 -> 5)) in 0 seconds Nov 27 04:55:38 localhost podman[242678]: time="2025-11-27T09:55:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:55:38 localhost podman[242678]: @ - - [27/Nov/2025:09:55:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:55:38 localhost podman[242678]: @ - - [27/Nov/2025:09:55:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19705 "" "Go-http-client/1.1" Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 0742c5ec-9810-4460-a29b-7709bfb18f26 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 0742c5ec-9810-4460-a29b-7709bfb18f26 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:38 localhost ceph-mgr[290377]: [progress INFO root] Completed event 0742c5ec-9810-4460-a29b-7709bfb18f26 (Updating node-proxy deployment (+5 -> 5)) in 0 seconds Nov 27 04:55:39 localhost nova_compute[284026]: 2025-11-27 09:55:39.235 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:39 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:39 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:39 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:39 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:39 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: Health check failed: 1 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:55:39 localhost ceph-mon[291772]: Health check failed: 1 stray host(s) with 1 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:39 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:39 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #25. Immutable memtables: 0. Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.273646) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 11] Flushing memtable with next log file: 25 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340273689, "job": 11, "event": "flush_started", "num_memtables": 1, "num_entries": 1335, "num_deletes": 252, "total_data_size": 6270921, "memory_usage": 6825296, "flush_reason": "Manual Compaction"} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 11] Level-0 flush table #26: started Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340309910, "cf_name": "default", "job": 11, "event": "table_file_creation", "file_number": 26, "file_size": 3970395, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 15843, "largest_seqno": 17173, "table_properties": {"data_size": 3963998, "index_size": 3414, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1925, "raw_key_size": 16944, "raw_average_key_size": 22, "raw_value_size": 3950137, "raw_average_value_size": 5273, "num_data_blocks": 143, "num_entries": 749, "num_filter_entries": 749, "num_deletions": 251, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237321, "oldest_key_time": 1764237321, "file_creation_time": 1764237340, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 26, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 11] Flush lasted 36334 microseconds, and 8361 cpu microseconds. Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.309977) [db/flush_job.cc:967] [default] [JOB 11] Level-0 flush table #26: 3970395 bytes OK Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.310005) [db/memtable_list.cc:519] [default] Level-0 commit table #26 started Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.312710) [db/memtable_list.cc:722] [default] Level-0 commit table #26: memtable #1 done Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.312733) EVENT_LOG_v1 {"time_micros": 1764237340312727, "job": 11, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.312756) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 11] Try to delete WAL files size 6263868, prev total WAL file size 6263868, number of live WAL files 2. Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000022.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.313925) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003130353432' seq:72057594037927935, type:22 .. '7061786F73003130373934' seq:0, type:0; will stop at (end) Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 12] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 11 Base level 0, inputs: [26(3877KB)], [24(15MB)] Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340313999, "job": 12, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [26], "files_L6": [24], "score": -1, "input_data_size": 19938380, "oldest_snapshot_seqno": -1} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 12] Generated table #27: 10191 keys, 16462205 bytes, temperature: kUnknown Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340440094, "cf_name": "default", "job": 12, "event": "table_file_creation", "file_number": 27, "file_size": 16462205, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 16403181, "index_size": 32427, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 25541, "raw_key_size": 274840, "raw_average_key_size": 26, "raw_value_size": 16227952, "raw_average_value_size": 1592, "num_data_blocks": 1227, "num_entries": 10191, "num_filter_entries": 10191, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237340, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 27, "seqno_to_time_mapping": "N/A"}} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.440423) [db/compaction/compaction_job.cc:1663] [default] [JOB 12] Compacted 1@0 + 1@6 files to L6 => 16462205 bytes Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.442100) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 158.0 rd, 130.4 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(3.8, 15.2 +0.0 blob) out(15.7 +0.0 blob), read-write-amplify(9.2) write-amplify(4.1) OK, records in: 10730, records dropped: 539 output_compression: NoCompression Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.442130) EVENT_LOG_v1 {"time_micros": 1764237340442117, "job": 12, "event": "compaction_finished", "compaction_time_micros": 126220, "compaction_time_cpu_micros": 28504, "output_level": 6, "num_output_files": 1, "total_output_size": 16462205, "num_input_records": 10730, "num_output_records": 10191, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000026.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340442774, "job": 12, "event": "table_file_deletion", "file_number": 26} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000024.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237340445213, "job": 12, "event": "table_file_deletion", "file_number": 24} Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.313796) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.445351) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.445358) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.445362) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.445365) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:55:40.445368) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:55:40 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v8: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 27 KiB/s rd, 0 B/s wr, 15 op/s Nov 27 04:55:41 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 04:55:41 localhost nova_compute[284026]: 2025-11-27 09:55:41.992 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:42 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:42 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v9: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 22 KiB/s rd, 0 B/s wr, 12 op/s Nov 27 04:55:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34437 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""], "format": "json"}]: dispatch Nov 27 04:55:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:55:43.565 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:55:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:55:43.565 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:55:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:55:43.566 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.136 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.137 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.137 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.150 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.151 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '72c866ae-7830-4d91-b135-c20cf424d18d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.137874', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d7b1c6e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': '529a11c91ffae65039da977034aaf2dd1ecfff32cdae1d240777bd7b38f3950a'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.137874', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d7b3424-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': '561db074773a9ef5e852e8ec623361bc481b16201a5297d1bccf6e1f0189e7f7'}]}, 'timestamp': '2025-11-27 09:55:44.151876', '_unique_id': '3ab51205b91a4a1ba22abdbe21ad6912'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.153 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.155 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.158 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c78aefd0-c9ac-4cf8-9c3b-c92cba717adc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.155140', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7c552a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'abbd2a90ca919ed41756a2188da58c9f75c80dc066a27dfbe05177e751b833e8'}]}, 'timestamp': '2025-11-27 09:55:44.159295', '_unique_id': 'aa5874c0f1b147519895de29ed78c59c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.160 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.161 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.161 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5a709824-e079-4778-a32e-eed769e5844c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.161456', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7cbd3a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': '4497f3ea4aa4f9a964e5b60c5fcd2b957973e0e152cd1c54eda39ffb433fcc98'}]}, 'timestamp': '2025-11-27 09:55:44.161954', '_unique_id': '9ff1a7e377f74a5f884cdbeb6453b861'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.162 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.163 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.164 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f3288fc3-577c-4bb0-894a-b8e27e9d453b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.164024', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7d1faa-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'ff3fa3f635ac621f38641e449c40275066dbf393ef276f282e5bd5bc681a80da'}]}, 'timestamp': '2025-11-27 09:55:44.164466', '_unique_id': '3d615e0dc840458f8c507aa45f6429d3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.165 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.166 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '85651762-3d72-435b-8b98-df0b15bb8c08', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.166562', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7d82f6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'c32b803b286b6cbcd12a12d9f545535323692c5291ebc4ba13e538a6bdc5c249'}]}, 'timestamp': '2025-11-27 09:55:44.167010', '_unique_id': '368e5837aa994f9db530d3680743dd80'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.168 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eb401fa6-8c6a-41a3-807e-b19d2fff45b9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.169060', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7de44e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'b91bf3ef0675665175b64715f0b6a7aef328b32d8446db8f0f73ab27f71e3d19'}]}, 'timestamp': '2025-11-27 09:55:44.169500', '_unique_id': '9f5dc3c850314ecc833c7d62a51e55e3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.171 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.171 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'cb68ab70-6204-461b-8dcd-3a883476aae6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.171553', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d7e4678-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'c4782cd998efcbd2b011ff87aa194b4a59aeec352c00a38fc5551f44ed1a609c'}]}, 'timestamp': '2025-11-27 09:55:44.172013', '_unique_id': '5ea0649d7af347f6adca1a820491851e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.173 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.174 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.200 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.200 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd4c748b1-3fa8-4738-9b49-d895c7fce4ef', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.174235', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d82ac68-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '2cd4ff4aaee6ed751f259cc06718281a369cd02a39328c381474f60987cd3742'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.174235', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d82c0d6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': 'fe0ac9dc9d9a68a581a513f7898262e4a94cc07d71f59d3e1a98f92e1d8e92c3'}]}, 'timestamp': '2025-11-27 09:55:44.201353', '_unique_id': '37e83f68f7f748459005ee39c3f50047'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.203 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.204 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.204 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '85908e24-0d30-49a1-b27e-bd8521d0477b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.204120', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d833e9e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '7779f992d0d45a0a2f2d213de9db7591c0920f210953faf359bddf808fc8fbdb'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.204120', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d834fa6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '079c34099216c86fc49488b107b2284df96db3d515c170d44418c0b6247a109f'}]}, 'timestamp': '2025-11-27 09:55:44.205013', '_unique_id': '9c19338be60b4853bc7ef3ee1ac968e0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.207 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.207 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.207 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.207 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '22461ed1-bf3f-4fcd-941a-3f4f264ac8a1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.207299', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d83bb30-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': 'c23429ccd34fceeb335ba71d7fb443ad25111485f0527569e3f38d4845519b62'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.207299', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d83cae4-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '30e556b639417d92f9c53c2f5edc97b21816958bdbb1cc663f63c2ffddb7f51f'}]}, 'timestamp': '2025-11-27 09:55:44.208143', '_unique_id': 'e40300a6b75d4e7bb30199bf762e991b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.210 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.210 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e357b72f-f911-42af-879d-c7a7f9f4820b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.210274', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d842f02-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': 'ee1a3806a2ab6000e237d07413e8f508ac61d2b8a00be938c53d0fbe8bc72d68'}]}, 'timestamp': '2025-11-27 09:55:44.210767', '_unique_id': 'febf1a813d994383a84826f5afbc3c12'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.212 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5a6cdf6c-0c9c-4f93-b992-2f7ea6f16813', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:55:44.212816', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '3d86af48-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.414449496, 'message_signature': '38ee40f880a77e7c87c78145715545875e82d564518fa16dc91db7e47219d43b'}]}, 'timestamp': '2025-11-27 09:55:44.227213', '_unique_id': '9b5c51b7476a47e1b791f2d465d34147'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.230 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '360adad5-89c7-43a0-a2d3-473c7df4abf7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.230347', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d87428c-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': 'fcd49d0ebe0e1b0d7f9d7b0aef15020316a6ffa478ef99db6f064da2f04d58ac'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.230347', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d875862-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '95351ebd47ad11282f1ddcb0e5dae634e93fc2c924154298923dd098edb59926'}]}, 'timestamp': '2025-11-27 09:55:44.231560', '_unique_id': '71ee5dd0f60d4f06a11b817fc0271da3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.234 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.234 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceph-mon[291772]: mon.np0005537446@2(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:44 localhost nova_compute[284026]: 2025-11-27 09:55:44.282 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e429454a-20ab-45a0-b8a8-bc450b38494a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.234339', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d87db84-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': 'f0e207d39731a7ec0ef3d8247c10ef63812aa595a8a0097e4ad9b1355834bac6'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.234339', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d8ef6ee-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': '1c36d469ae6de1f1006d96eb585eaab943ac1fd8b047e178727b6c8d21e37a1d'}]}, 'timestamp': '2025-11-27 09:55:44.281474', '_unique_id': '1fabf7cf49f34158b6bb954ab2b6ae7c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.283 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.283 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 13030000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '93185ae5-2044-4a0f-8e2d-67ed90a692ae', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 13030000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:55:44.283361', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '3d8f527e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.414449496, 'message_signature': '458305c70265b19bfd84ede99edf83460299664205d1a8bed1fdca3310b66294'}]}, 'timestamp': '2025-11-27 09:55:44.283635', '_unique_id': 'ad2d3bb8425f4c808cfb10cc482ae24a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.284 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5cfdf653-f5cf-48c6-8ad2-41a409ba99fe', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.284630', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d8f82b2-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': '6e3209fab8e344fd39e7aead1691e01aead90c529854bfd58c194b4a711bb312'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.284630', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d8f8ac8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.325930531, 'message_signature': 'ba2d0cae1cdf4172e2138ccf792041555a34a11ebaa53fe4c4efaef4d87d6a34'}]}, 'timestamp': '2025-11-27 09:55:44.285042', '_unique_id': '102d04805ba247a0a9c56d12d776c62e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.285 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1417b32e-dd70-4877-a9f6-6f334b7cc2d4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.286033', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d8fb9b2-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': '774fd0970344d5256ad8d1d59ba465e626826ea1b17efb5d6abd743dbe8b4a7b'}]}, 'timestamp': '2025-11-27 09:55:44.286257', '_unique_id': 'c372112eee36427ab26bdda6f0f0da07'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.286 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e8f44128-c4a6-47e9-ad77-0a8091192f76', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.287286', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d8fea18-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': '97fbe9b3d4c6dcd46e0dad5fc6c564adeecdd77ceae8b17f47ec9ca4f864f1cc'}]}, 'timestamp': '2025-11-27 09:55:44.287494', '_unique_id': '584163cc39e949439e3697706b4d4cf1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.288 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.288 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4545ce3c-f3b4-4cae-aabb-3145c64edbfd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:55:44.288857', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3d90285c-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.343221034, 'message_signature': '09ef1dd2028cdf9172769af66866ab0bd9bed1ff0951ababfff7c2fa7df99826'}]}, 'timestamp': '2025-11-27 09:55:44.289090', '_unique_id': '2ca46bef5df54a5f9d5ae643bc9e1434'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.289 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '507cdf2c-92fb-4e2a-b716-fe25e318b768', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.290049', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d905624-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '831f4d9cc1b4b95c408f5cc451929d824b0ed21a280586cbf6d6d4ac36fcafca'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.290049', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d905d54-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '2d7d3c435f5c9a34a96a95e6c3389464f0597c772c2b1d18cce9af0781da5898'}]}, 'timestamp': '2025-11-27 09:55:44.290430', '_unique_id': 'cd31c4d25ac44f74932bda7630f4d8c5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.290 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.291 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.291 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.291 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6b72a6b6-c3e4-44d7-bc59-5f134244a849', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:55:44.291408', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3d908b3a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': 'a3a61a3ebd19002d2c96df14cdbca69e732306ca50fd846b06df64b7828a9f2e'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:55:44.291408', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3d90931e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11574.362304316, 'message_signature': '737c8b01b08947bdfc6298d70b0b38fe1a4542b36a04d6ef5438d2e52224212a'}]}, 'timestamp': '2025-11-27 09:55:44.291806', '_unique_id': '695817fd4d7649f49b7720627f6bb743'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:55:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:55:44.292 12 ERROR oslo_messaging.notify.messaging Nov 27 04:55:44 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v10: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 04:55:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:55:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:55:44 localhost systemd[1]: tmp-crun.UU4UXV.mount: Deactivated successfully. Nov 27 04:55:44 localhost podman[299658]: 2025-11-27 09:55:44.726859897 +0000 UTC m=+0.091044553 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 04:55:44 localhost podman[299658]: 2025-11-27 09:55:44.735964262 +0000 UTC m=+0.100148938 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:55:44 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:55:44 localhost podman[299659]: 2025-11-27 09:55:44.788032258 +0000 UTC m=+0.148429673 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd) Nov 27 04:55:44 localhost podman[299659]: 2025-11-27 09:55:44.806919395 +0000 UTC m=+0.167316790 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd) Nov 27 04:55:44 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:55:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.27055 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "mon", "daemon_id": "np0005537442", "target": ["mon-mgr", ""], "format": "json"}]: dispatch Nov 27 04:55:46 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v11: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 04:55:47 localhost nova_compute[284026]: 2025-11-27 09:55:47.035 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34398 -' entity='client.admin' cmd=[{"prefix": "orch daemon rm", "names": ["mon.np0005537442"], "force": true, "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:55:47 localhost ceph-mgr[290377]: [cephadm INFO root] Remove daemons mon.np0005537442 Nov 27 04:55:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Remove daemons mon.np0005537442 Nov 27 04:55:47 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Safe to remove mon.np0005537442: new quorum should be ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444'] (from ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444']) Nov 27 04:55:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Safe to remove mon.np0005537442: new quorum should be ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444'] (from ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444']) Nov 27 04:55:47 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Removing monitor np0005537442 from monmap... Nov 27 04:55:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing monitor np0005537442 from monmap... Nov 27 04:55:47 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing daemon mon.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:55:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing daemon mon.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:55:47 localhost ceph-mon[291772]: mon.np0005537446@2(peon) e10 my rank is now 1 (was 2) Nov 27 04:55:47 localhost ceph-mgr[290377]: client.34422 ms_handle_reset on v2:172.18.0.108:3300/0 Nov 27 04:55:47 localhost ceph-mgr[290377]: client.34427 ms_handle_reset on v2:172.18.0.108:3300/0 Nov 27 04:55:47 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:55:47 localhost ceph-mon[291772]: paxos.1).electionLogic(48) init, last seen epoch 48 Nov 27 04:55:47 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e10 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:55:47 localhost ceph-osd[32722]: --2- [v2:172.18.0.108:6804/1946346282,v1:172.18.0.108:6805/1946346282] >> [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] conn(0x5568f6fbd800 0x5568fba4e580 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=1 rev1=1 crypto rx=0 tx=0 comp rx=0 tx=0).send_auth_request get_initial_auth_request returned -2 Nov 27 04:55:47 localhost ceph-mgr[290377]: --2- 172.18.0.108:0/2668738776 >> [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] conn(0x556daaa37400 0x556daaf57700 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=1 rev1=1 crypto rx=0 tx=0 comp rx=0 tx=0).send_auth_request get_initial_auth_request returned -2 Nov 27 04:55:47 localhost ceph-mds[289027]: --2- [v2:172.18.0.108:6808/2569123203,v1:172.18.0.108:6809/2569123203] >> [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] conn(0x5582255ff800 0x55822555a680 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=1 rev1=1 crypto rx=0 tx=0 comp rx=0 tx=0).send_auth_request get_initial_auth_request returned -2 Nov 27 04:55:48 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v12: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 04:55:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:55:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:55:49 localhost podman[299697]: 2025-11-27 09:55:48.998874843 +0000 UTC m=+0.088601828 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, managed_by=edpm_ansible, vcs-type=git, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.component=ubi9-minimal-container, architecture=x86_64, config_id=edpm, distribution-scope=public, version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.expose-services=, container_name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, build-date=2025-08-20T13:12:41, maintainer=Red Hat, Inc., io.buildah.version=1.33.7, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9) Nov 27 04:55:49 localhost podman[299697]: 2025-11-27 09:55:49.044668902 +0000 UTC m=+0.134395917 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, build-date=2025-08-20T13:12:41, vcs-type=git, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., config_id=edpm, release=1755695350, com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., version=9.6, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 04:55:49 localhost podman[299696]: 2025-11-27 09:55:49.054550206 +0000 UTC m=+0.148209627 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:55:49 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:55:49 localhost podman[299696]: 2025-11-27 09:55:49.068899761 +0000 UTC m=+0.162559162 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:55:49 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:55:49 localhost nova_compute[284026]: 2025-11-27 09:55:49.293 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e10 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e10 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:55:49 localhost ceph-mon[291772]: Safe to remove mon.np0005537442: new quorum should be ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444'] (from ['np0005537443', 'np0005537446', 'np0005537445', 'np0005537444']) Nov 27 04:55:49 localhost ceph-mon[291772]: Removing monitor np0005537442 from monmap... Nov 27 04:55:49 localhost ceph-mon[291772]: Removing daemon mon.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:55:49 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537446,np0005537445,np0005537444 in quorum (ranks 0,1,2,3) Nov 27 04:55:49 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 04:55:49 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm Nov 27 04:55:49 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:55:49 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 04:55:49 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.44339 -' entity='client.admin' cmd=[{"prefix": "orch host label rm", "hostname": "np0005537442.localdomain", "label": "mon", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO root] Removed label mon from host np0005537442.localdomain Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removed label mon from host np0005537442.localdomain Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v13: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:50 localhost ceph-mon[291772]: Removed label mon from host np0005537442.localdomain Nov 27 04:55:50 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:50 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.27080 -' entity='client.admin' cmd=[{"prefix": "orch host label rm", "hostname": "np0005537442.localdomain", "label": "mgr", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:55:51 localhost ceph-mgr[290377]: [cephadm INFO root] Removed label mgr from host np0005537442.localdomain Nov 27 04:55:51 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removed label mgr from host np0005537442.localdomain Nov 27 04:55:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:55:51 localhost podman[300019]: 2025-11-27 09:55:51.972107787 +0000 UTC m=+0.084911960 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:55:51 localhost podman[300019]: 2025-11-27 09:55:51.981689193 +0000 UTC m=+0.094493316 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:55:52 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:55:52 localhost nova_compute[284026]: 2025-11-27 09:55:52.067 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:52 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev dc936adf-e1d3-46f6-802a-0a6065032d0c (Updating mgr deployment (-1 -> 4)) Nov 27 04:55:52 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing daemon mgr.np0005537442.byetac from np0005537442.localdomain -- ports [8765] Nov 27 04:55:52 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing daemon mgr.np0005537442.byetac from np0005537442.localdomain -- ports [8765] Nov 27 04:55:52 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v14: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:52 localhost ceph-mon[291772]: Removed label mgr from host np0005537442.localdomain Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:52 localhost ceph-mon[291772]: Removing daemon mgr.np0005537442.byetac from np0005537442.localdomain -- ports [8765] Nov 27 04:55:52 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:54 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:54 localhost nova_compute[284026]: 2025-11-27 09:55:54.327 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:54 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Removing key for mgr.np0005537442.byetac Nov 27 04:55:54 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing key for mgr.np0005537442.byetac Nov 27 04:55:54 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev dc936adf-e1d3-46f6-802a-0a6065032d0c (Updating mgr deployment (-1 -> 4)) Nov 27 04:55:54 localhost ceph-mgr[290377]: [progress INFO root] Completed event dc936adf-e1d3-46f6-802a-0a6065032d0c (Updating mgr deployment (-1 -> 4)) in 2 seconds Nov 27 04:55:54 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 6cae6d6e-7c1d-49c9-88a9-f1daadee6367 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:54 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 6cae6d6e-7c1d-49c9-88a9-f1daadee6367 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:54 localhost ceph-mgr[290377]: [progress INFO root] Completed event 6cae6d6e-7c1d-49c9-88a9-f1daadee6367 (Updating node-proxy deployment (+5 -> 5)) in 0 seconds Nov 27 04:55:54 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v15: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:55 localhost ceph-mon[291772]: Removing key for mgr.np0005537442.byetac Nov 27 04:55:55 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "mgr.np0005537442.byetac"} : dispatch Nov 27 04:55:55 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "mgr.np0005537442.byetac"}]': finished Nov 27 04:55:55 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:55 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:55 localhost openstack_network_exporter[244641]: ERROR 09:55:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:55:55 localhost openstack_network_exporter[244641]: ERROR 09:55:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:55:55 localhost openstack_network_exporter[244641]: ERROR 09:55:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:55:55 localhost openstack_network_exporter[244641]: ERROR 09:55:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:55:55 localhost openstack_network_exporter[244641]: Nov 27 04:55:55 localhost openstack_network_exporter[244641]: ERROR 09:55:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:55:55 localhost openstack_network_exporter[244641]: Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.44367 -' entity='client.admin' cmd=[{"prefix": "orch host label rm", "hostname": "np0005537442.localdomain", "label": "_admin", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:55:56 localhost ceph-mgr[290377]: [cephadm INFO root] Removed label _admin from host np0005537442.localdomain Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removed label _admin from host np0005537442.localdomain Nov 27 04:55:56 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:56 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:56 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 6301d42a-5b63-415b-b001-e1a346319d13 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 6301d42a-5b63-415b-b001-e1a346319d13 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] Completed event 6301d42a-5b63-415b-b001-e1a346319d13 (Updating node-proxy deployment (+5 -> 5)) in 0 seconds Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:56 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v16: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev e8ea74ec-51af-46df-b5ba-7baf2895a2d9 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev e8ea74ec-51af-46df-b5ba-7baf2895a2d9 (Updating node-proxy deployment (+5 -> 5)) Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] Completed event e8ea74ec-51af-46df-b5ba-7baf2895a2d9 (Updating node-proxy deployment (+5 -> 5)) in 0 seconds Nov 27 04:55:56 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 04:55:57 localhost nova_compute[284026]: 2025-11-27 09:55:57.108 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:55:57 localhost ceph-mon[291772]: Removed label _admin from host np0005537442.localdomain Nov 27 04:55:57 localhost ceph-mon[291772]: Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:55:57 localhost ceph-mon[291772]: Removing np0005537442.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:55:57 localhost ceph-mon[291772]: Removing np0005537442.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:55:57 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:55:57 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:57 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:55:58 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v17: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:55:59 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:55:59 localhost nova_compute[284026]: 2025-11-27 09:55:59.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:00 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v18: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:56:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:56:00 localhost podman[300133]: 2025-11-27 09:56:00.995762345 +0000 UTC m=+0.086155733 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:56:01 localhost podman[300133]: 2025-11-27 09:56:01.010578962 +0000 UTC m=+0.100972360 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:56:01 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:56:01 localhost podman[300132]: 2025-11-27 09:56:01.088833102 +0000 UTC m=+0.184418769 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 04:56:01 localhost podman[300132]: 2025-11-27 09:56:01.167089571 +0000 UTC m=+0.262675218 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 04:56:01 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:01 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:02 localhost nova_compute[284026]: 2025-11-27 09:56:02.110 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:02 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v19: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:04 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:04 localhost nova_compute[284026]: 2025-11-27 09:56:04.367 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:04 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v20: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:06 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v21: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:07 localhost nova_compute[284026]: 2025-11-27 09:56:07.149 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34418 -' entity='client.admin' cmd=[{"prefix": "orch host drain", "hostname": "np0005537442.localdomain", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:07 localhost ceph-mgr[290377]: [cephadm INFO root] Added label _no_schedule to host np0005537442.localdomain Nov 27 04:56:07 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Added label _no_schedule to host np0005537442.localdomain Nov 27 04:56:07 localhost ceph-mgr[290377]: [cephadm INFO root] Added label SpecialHostLabels.DRAIN_CONF_KEYRING to host np0005537442.localdomain Nov 27 04:56:07 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Added label SpecialHostLabels.DRAIN_CONF_KEYRING to host np0005537442.localdomain Nov 27 04:56:07 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev dca0d896-a18d-442f-b73e-f5f50214ac37 (Updating crash deployment (-1 -> 4)) Nov 27 04:56:07 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing daemon crash.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:56:07 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing daemon crash.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:56:07 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:07 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:56:07 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:07 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:08 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v22: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:08 localhost podman[242678]: time="2025-11-27T09:56:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:56:08 localhost podman[242678]: @ - - [27/Nov/2025:09:56:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:56:08 localhost podman[242678]: @ - - [27/Nov/2025:09:56:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19709 "" "Go-http-client/1.1" Nov 27 04:56:08 localhost ceph-mon[291772]: Added label _no_schedule to host np0005537442.localdomain Nov 27 04:56:08 localhost ceph-mon[291772]: Added label SpecialHostLabels.DRAIN_CONF_KEYRING to host np0005537442.localdomain Nov 27 04:56:08 localhost ceph-mon[291772]: Removing daemon crash.np0005537442 from np0005537442.localdomain -- ports [] Nov 27 04:56:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.27100 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "host_pattern": "np0005537442.localdomain", "target": ["mon-mgr", ""], "format": "json"}]: dispatch Nov 27 04:56:09 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:09 localhost nova_compute[284026]: 2025-11-27 09:56:09.399 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:10 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Removing key for client.crash.np0005537442.localdomain Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing key for client.crash.np0005537442.localdomain Nov 27 04:56:10 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev dca0d896-a18d-442f-b73e-f5f50214ac37 (Updating crash deployment (-1 -> 4)) Nov 27 04:56:10 localhost ceph-mgr[290377]: [progress INFO root] Completed event dca0d896-a18d-442f-b73e-f5f50214ac37 (Updating crash deployment (-1 -> 4)) in 2 seconds Nov 27 04:56:10 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 909397c9-147f-48b1-8a75-ea6a3bf1c634 (Updating node-proxy deployment (+4 -> 4)) Nov 27 04:56:10 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 909397c9-147f-48b1-8a75-ea6a3bf1c634 (Updating node-proxy deployment (+4 -> 4)) Nov 27 04:56:10 localhost ceph-mgr[290377]: [progress INFO root] Completed event 909397c9-147f-48b1-8a75-ea6a3bf1c634 (Updating node-proxy deployment (+4 -> 4)) in 0 seconds Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34477 -' entity='client.admin' cmd=[{"prefix": "orch host rm", "hostname": "np0005537442.localdomain", "force": true, "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:10 localhost ceph-mgr[290377]: [cephadm ERROR cephadm.utils] executing refresh((['np0005537442.localdomain', 'np0005537443.localdomain', 'np0005537444.localdomain', 'np0005537445.localdomain', 'np0005537446.localdomain'],)) failed.#012Traceback (most recent call last):#012 File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work#012 return f(*arg)#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh#012 if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance":#012KeyError: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(cephadm) log [ERR] : executing refresh((['np0005537442.localdomain', 'np0005537443.localdomain', 'np0005537444.localdomain', 'np0005537445.localdomain', 'np0005537446.localdomain'],)) failed.#012Traceback (most recent call last):#012 File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work#012 return f(*arg)#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh#012 if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance":#012KeyError: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:10.505+0000 7fae165cd640 -1 log_channel(cephadm) log [ERR] : executing refresh((['np0005537442.localdomain', 'np0005537443.localdomain', 'np0005537444.localdomain', 'np0005537445.localdomain', 'np0005537446.localdomain'],)) failed. Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Traceback (most recent call last): Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: return f(*arg) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance": Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: KeyError: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(cluster) log [ERR] : Unhandled exception from module 'cephadm' while running on mgr.np0005537446.nfulyg: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-mgr[290377]: cephadm.serve: Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:10.505+0000 7fae0f57f640 -1 log_channel(cluster) log [ERR] : Unhandled exception from module 'cephadm' while running on mgr.np0005537446.nfulyg: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:10.505+0000 7fae0f57f640 -1 cephadm.serve: Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v23: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:10 localhost ceph-mgr[290377]: Traceback (most recent call last):#012 File "/usr/share/ceph/mgr/cephadm/module.py", line 763, in serve#012 serve.serve()#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 82, in serve#012 self._refresh_hosts_and_daemons()#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 330, in _refresh_hosts_and_daemons#012 refresh(self.mgr.cache.get_hosts())#012 File "/usr/share/ceph/mgr/cephadm/utils.py", line 100, in forall_hosts_wrapper#012 return CephadmOrchestrator.instance._worker_pool.map(do_work, vals)#012 File "/lib64/python3.9/multiprocessing/pool.py", line 364, in map#012 return self._map_async(func, iterable, mapstar, chunksize).get()#012 File "/lib64/python3.9/multiprocessing/pool.py", line 771, in get#012 raise self._value#012 File "/lib64/python3.9/multiprocessing/pool.py", line 125, in worker#012 result = (True, func(*args, **kwds))#012 File "/lib64/python3.9/multiprocessing/pool.py", line 48, in mapstar#012 return list(map(*args))#012 File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work#012 return f(*arg)#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh#012 if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance":#012KeyError: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:10.505+0000 7fae0f57f640 -1 Traceback (most recent call last): Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/module.py", line 763, in serve Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: serve.serve() Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/serve.py", line 82, in serve Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: self._refresh_hosts_and_daemons() Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/serve.py", line 330, in _refresh_hosts_and_daemons Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: refresh(self.mgr.cache.get_hosts()) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/utils.py", line 100, in forall_hosts_wrapper Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: return CephadmOrchestrator.instance._worker_pool.map(do_work, vals) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/lib64/python3.9/multiprocessing/pool.py", line 364, in map Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: return self._map_async(func, iterable, mapstar, chunksize).get() Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/lib64/python3.9/multiprocessing/pool.py", line 771, in get Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: raise self._value Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/lib64/python3.9/multiprocessing/pool.py", line 125, in worker Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: result = (True, func(*args, **kwds)) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/lib64/python3.9/multiprocessing/pool.py", line 48, in mapstar Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: return list(map(*args)) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: return f(*arg) Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance": Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: KeyError: 'np0005537442.localdomain' Nov 27 04:56:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Nov 27 04:56:10 localhost ceph-mgr[290377]: [cephadm INFO root] Removed host np0005537442.localdomain Nov 27 04:56:10 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removed host np0005537442.localdomain Nov 27 04:56:10 localhost ceph-mon[291772]: Removing key for client.crash.np0005537442.localdomain Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.crash.np0005537442.localdomain"} : dispatch Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.crash.np0005537442.localdomain"}]': finished Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain"} : dispatch Nov 27 04:56:10 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain"}]': finished Nov 27 04:56:11 localhost ceph-mon[291772]: executing refresh((['np0005537442.localdomain', 'np0005537443.localdomain', 'np0005537444.localdomain', 'np0005537445.localdomain', 'np0005537446.localdomain'],)) failed.#012Traceback (most recent call last):#012 File "/usr/share/ceph/mgr/cephadm/utils.py", line 94, in do_work#012 return f(*arg)#012 File "/usr/share/ceph/mgr/cephadm/serve.py", line 252, in refresh#012 if self.mgr.inventory._inventory[host].get("status", "").lower() == "maintenance":#012KeyError: 'np0005537442.localdomain' Nov 27 04:56:11 localhost ceph-mon[291772]: Unhandled exception from module 'cephadm' while running on mgr.np0005537446.nfulyg: 'np0005537442.localdomain' Nov 27 04:56:11 localhost ceph-mon[291772]: Removed host np0005537442.localdomain Nov 27 04:56:11 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 04:56:12 localhost nova_compute[284026]: 2025-11-27 09:56:12.185 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:12 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v24: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:12 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:13 localhost nova_compute[284026]: 2025-11-27 09:56:13.086 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:13 localhost nova_compute[284026]: 2025-11-27 09:56:13.087 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:13 localhost nova_compute[284026]: 2025-11-27 09:56:13.087 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:13 localhost nova_compute[284026]: 2025-11-27 09:56:13.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:13 localhost nova_compute[284026]: 2025-11-27 09:56:13.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:56:13 localhost ceph-mon[291772]: Health check failed: Module 'cephadm' has failed: 'np0005537442.localdomain' (MGR_MODULE_ERROR) Nov 27 04:56:14 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:14 localhost nova_compute[284026]: 2025-11-27 09:56:14.422 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:14 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v25: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:14 localhost nova_compute[284026]: 2025-11-27 09:56:14.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:14 localhost nova_compute[284026]: 2025-11-27 09:56:14.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:56:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:56:14 localhost podman[300200]: 2025-11-27 09:56:14.975563523 +0000 UTC m=+0.067000598 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd) Nov 27 04:56:14 localhost podman[300200]: 2025-11-27 09:56:14.992888558 +0000 UTC m=+0.084325703 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 04:56:15 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:56:15 localhost systemd[1]: tmp-crun.ErNqFi.mount: Deactivated successfully. Nov 27 04:56:15 localhost podman[300199]: 2025-11-27 09:56:15.0406691 +0000 UTC m=+0.134348166 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent) Nov 27 04:56:15 localhost podman[300199]: 2025-11-27 09:56:15.075939306 +0000 UTC m=+0.169618372 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:56:15 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.737 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.881 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.882 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.882 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.883 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:56:15 localhost nova_compute[284026]: 2025-11-27 09:56:15.884 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:56:16 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e10 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:56:16 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3792291855' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.331 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.447s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.468 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.469 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:56:16 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v26: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:16 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e10 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:56:16 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.106:0/131418844' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.712 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.714 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11458MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.715 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.716 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.931 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.932 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.932 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:56:16 localhost nova_compute[284026]: 2025-11-27 09:56:16.975 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.231 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.456 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.481s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.464 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.510 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.512 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:56:17 localhost nova_compute[284026]: 2025-11-27 09:56:17.513 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.797s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:56:18 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34442 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:18 localhost ceph-mgr[290377]: [cephadm INFO root] Saving service mon spec with placement label:mon Nov 27 04:56:18 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Saving service mon spec with placement label:mon Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.509 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.510 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.510 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:56:18 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v27: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.882 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.883 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.883 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:56:18 localhost nova_compute[284026]: 2025-11-27 09:56:18.883 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:56:19 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:56:19 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:19 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:19 localhost nova_compute[284026]: 2025-11-27 09:56:19.373 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:56:19 localhost nova_compute[284026]: 2025-11-27 09:56:19.410 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:56:19 localhost nova_compute[284026]: 2025-11-27 09:56:19.411 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:56:19 localhost nova_compute[284026]: 2025-11-27 09:56:19.456 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.27140 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "mon", "daemon_id": "np0005537445", "target": ["mon-mgr", ""], "format": "json"}]: dispatch Nov 27 04:56:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:56:19 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:56:19 localhost podman[300278]: 2025-11-27 09:56:19.996778358 +0000 UTC m=+0.081050716 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute) Nov 27 04:56:20 localhost systemd[1]: tmp-crun.xWVR70.mount: Deactivated successfully. Nov 27 04:56:20 localhost podman[300279]: 2025-11-27 09:56:20.062755467 +0000 UTC m=+0.141654970 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, io.buildah.version=1.33.7, release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, config_id=edpm, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, vcs-type=git, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc.) Nov 27 04:56:20 localhost podman[300279]: 2025-11-27 09:56:20.075686925 +0000 UTC m=+0.154586498 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, build-date=2025-08-20T13:12:41, version=9.6, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, maintainer=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, distribution-scope=public, release=1755695350, name=ubi9-minimal, io.openshift.expose-services=) Nov 27 04:56:20 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:56:20 localhost podman[300278]: 2025-11-27 09:56:20.133747672 +0000 UTC m=+0.218020020 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 04:56:20 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:56:20 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v28: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:21 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.34513 -' entity='client.admin' cmd=[{"prefix": "orch daemon rm", "names": ["mon.np0005537445"], "force": true, "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:21 localhost ceph-mgr[290377]: [cephadm INFO root] Remove daemons mon.np0005537445 Nov 27 04:56:21 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Remove daemons mon.np0005537445 Nov 27 04:56:21 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Safe to remove mon.np0005537445: new quorum should be ['np0005537443', 'np0005537446', 'np0005537444'] (from ['np0005537443', 'np0005537446', 'np0005537444']) Nov 27 04:56:21 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Safe to remove mon.np0005537445: new quorum should be ['np0005537443', 'np0005537446', 'np0005537444'] (from ['np0005537443', 'np0005537446', 'np0005537444']) Nov 27 04:56:21 localhost ceph-mgr[290377]: [cephadm INFO cephadm.services.cephadmservice] Removing monitor np0005537445 from monmap... Nov 27 04:56:21 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing monitor np0005537445 from monmap... Nov 27 04:56:21 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Removing daemon mon.np0005537445 from np0005537445.localdomain -- ports [] Nov 27 04:56:21 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Removing daemon mon.np0005537445 from np0005537445.localdomain -- ports [] Nov 27 04:56:21 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:56:21 localhost ceph-mon[291772]: paxos.1).electionLogic(50) init, last seen epoch 50 Nov 27 04:56:21 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:21 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:22 localhost nova_compute[284026]: 2025-11-27 09:56:22.266 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:22 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v29: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:56:22 localhost podman[300318]: 2025-11-27 09:56:22.995458013 +0000 UTC m=+0.089275626 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:56:23 localhost podman[300318]: 2025-11-27 09:56:23.001070634 +0000 UTC m=+0.094888297 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:56:23 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:56:24 localhost nova_compute[284026]: 2025-11-27 09:56:24.492 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:24 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v30: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:25 localhost openstack_network_exporter[244641]: ERROR 09:56:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:56:25 localhost openstack_network_exporter[244641]: ERROR 09:56:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:56:25 localhost openstack_network_exporter[244641]: ERROR 09:56:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:56:25 localhost openstack_network_exporter[244641]: ERROR 09:56:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:56:25 localhost openstack_network_exporter[244641]: Nov 27 04:56:25 localhost openstack_network_exporter[244641]: ERROR 09:56:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:56:25 localhost openstack_network_exporter[244641]: Nov 27 04:56:26 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:26 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:56:26 localhost ceph-mon[291772]: paxos.1).electionLogic(53) init, last seen epoch 53, mid-election, bumping Nov 27 04:56:26 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:26 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e11 handle_timecheck drop unexpected msg Nov 27 04:56:26 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:26 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e11 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:26 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v31: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:27 localhost nova_compute[284026]: 2025-11-27 09:56:27.308 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:27 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:56:27 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:56:27 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:56:27 localhost ceph-mon[291772]: Health check failed: 1/3 mons down, quorum np0005537443,np0005537446 (MON_DOWN) Nov 27 04:56:27 localhost ceph-mon[291772]: Health detail: HEALTH_ERR 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm; Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:27 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:56:27 localhost ceph-mon[291772]: [ERR] MGR_MODULE_ERROR: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:27 localhost ceph-mon[291772]: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:27 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:56:27 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537446,np0005537444 in quorum (ranks 0,1,2) Nov 27 04:56:27 localhost ceph-mon[291772]: Health check cleared: MON_DOWN (was: 1/3 mons down, quorum np0005537443,np0005537446) Nov 27 04:56:27 localhost ceph-mon[291772]: Health detail: HEALTH_ERR 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm; Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:27 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 04:56:27 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:56:27 localhost ceph-mon[291772]: [ERR] MGR_MODULE_ERROR: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:27 localhost ceph-mon[291772]: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:28 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v32: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:29 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:29 localhost nova_compute[284026]: 2025-11-27 09:56:29.534 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:30 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v33: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:31 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_09:56:31 Nov 27 04:56:31 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 04:56:31 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 04:56:31 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'images', 'manila_metadata', '.mgr', 'manila_data', 'volumes', 'vms'] Nov 27 04:56:31 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 04:56:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:56:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.0014449417225013959 of space, bias 1.0, pg target 0.2885066972594454 quantized to 32 (current 32) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 04:56:31 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.453674623115578e-06 of space, bias 4.0, pg target 0.0019596681323283084 quantized to 16 (current 16) Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 04:56:31 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:56:31 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:56:32 localhost podman[300343]: 2025-11-27 09:56:32.019035269 +0000 UTC m=+0.108942433 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:56:32 localhost podman[300343]: 2025-11-27 09:56:32.082097262 +0000 UTC m=+0.172004446 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:56:32 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:56:32 localhost podman[300342]: 2025-11-27 09:56:31.995511449 +0000 UTC m=+0.089213445 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, io.buildah.version=1.41.3, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 04:56:32 localhost podman[300342]: 2025-11-27 09:56:32.146965052 +0000 UTC m=+0.240667048 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, config_id=ovn_controller, io.buildah.version=1.41.3, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 04:56:32 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:56:32 localhost nova_compute[284026]: 2025-11-27 09:56:32.309 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:32 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v34: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:34 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.44426 -' entity='client.admin' cmd=[{"prefix": "orch daemon add", "daemon_type": "mon", "placement": "np0005537445.localdomain:172.18.0.104", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:34 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Deploying daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:56:34 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Deploying daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:56:34 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v35: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:34 localhost nova_compute[284026]: 2025-11-27 09:56:34.560 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:35 localhost ceph-mon[291772]: Deploying daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:56:35 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:35 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:56:36 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v36: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:36 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e11 adding peer [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] to list of hints Nov 27 04:56:36 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e11 adding peer [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] to list of hints Nov 27 04:56:37 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e11 adding peer [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] to list of hints Nov 27 04:56:37 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:37 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (2) No such file or directory Nov 27 04:56:37 localhost nova_compute[284026]: 2025-11-27 09:56:37.349 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:37 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:37 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:38 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:38 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (2) No such file or directory Nov 27 04:56:38 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:38 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:56:38 localhost ceph-mon[291772]: paxos.1).electionLogic(56) init, last seen epoch 56 Nov 27 04:56:38 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e12 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:38 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v37: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:38 localhost podman[242678]: time="2025-11-27T09:56:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:56:38 localhost podman[242678]: @ - - [27/Nov/2025:09:56:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:56:38 localhost podman[242678]: @ - - [27/Nov/2025:09:56:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19715 "" "Go-http-client/1.1" Nov 27 04:56:39 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:39 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:39 localhost nova_compute[284026]: 2025-11-27 09:56:39.603 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:40 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:40 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:40 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v38: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:41 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:41 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:42 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:42 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:42 localhost ceph-mds[289027]: mds.beacon.mds.np0005537446.hkzfou missed beacon ack from the monitors Nov 27 04:56:42 localhost nova_compute[284026]: 2025-11-27 09:56:42.384 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:42 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v39: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:43 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:43 localhost ceph-mgr[290377]: mgr finish mon failed to return metadata for mon.np0005537445: (22) Invalid argument Nov 27 04:56:43 localhost ceph-mon[291772]: paxos.1).electionLogic(57) init, last seen epoch 57, mid-election, bumping Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e12 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537446@1(electing) e12 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e12 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:56:43 localhost ceph-mon[291772]: mon.np0005537443 is new leader, mons np0005537443,np0005537446,np0005537444,np0005537445 in quorum (ranks 0,1,2,3) Nov 27 04:56:43 localhost ceph-mon[291772]: Health detail: HEALTH_ERR 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm; Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:43 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm Nov 27 04:56:43 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:56:43 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 04:56:43 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:56:43 localhost ceph-mon[291772]: [ERR] MGR_MODULE_ERROR: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:43 localhost ceph-mon[291772]: Module 'cephadm' has failed: 'np0005537442.localdomain' Nov 27 04:56:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:56:43.566 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:56:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:56:43.566 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:56:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:56:43.567 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:56:44 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mon.np0005537445 172.18.0.107:0/3477689013; not ready for session (expect reconnect) Nov 27 04:56:44 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:44 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v40: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:44 localhost nova_compute[284026]: 2025-11-27 09:56:44.638 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:45 localhost ceph-mgr[290377]: mgr.server handle_report got status from non-daemon mon.np0005537445 Nov 27 04:56:45 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:45.202+0000 7fae1d61b640 -1 mgr.server handle_report got status from non-daemon mon.np0005537445 Nov 27 04:56:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:56:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:56:45 localhost systemd[1]: tmp-crun.VhKYoJ.mount: Deactivated successfully. Nov 27 04:56:46 localhost podman[300391]: 2025-11-27 09:56:46.003225775 +0000 UTC m=+0.093883060 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:56:46 localhost podman[300392]: 2025-11-27 09:56:46.054571062 +0000 UTC m=+0.144276651 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:56:46 localhost podman[300392]: 2025-11-27 09:56:46.061863878 +0000 UTC m=+0.151569467 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:56:46 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:56:46 localhost podman[300391]: 2025-11-27 09:56:46.11636257 +0000 UTC m=+0.207019855 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS) Nov 27 04:56:46 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:56:46 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v41: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:47 localhost nova_compute[284026]: 2025-11-27 09:56:47.387 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:48 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v42: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:49 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e86 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:49 localhost nova_compute[284026]: 2025-11-27 09:56:49.667 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:50 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v43: 177 pgs: 177 active+clean; 104 MiB data, 579 MiB used, 41 GiB / 42 GiB avail Nov 27 04:56:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.44450 -' entity='client.admin' cmd=[{"prefix": "orch", "action": "reconfig", "service_name": "osd.default_drive_group", "target": ["mon-mgr", ""]}]: dispatch Nov 27 04:56:50 localhost ceph-mgr[290377]: [cephadm INFO root] Reconfig service osd.default_drive_group Nov 27 04:56:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Reconfig service osd.default_drive_group Nov 27 04:56:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:56:50 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:56:51 localhost podman[300430]: 2025-11-27 09:56:51.008912583 +0000 UTC m=+0.091485426 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, tcib_managed=true, config_id=edpm, org.label-schema.build-date=20251125) Nov 27 04:56:51 localhost podman[300430]: 2025-11-27 09:56:51.046989454 +0000 UTC m=+0.129562247 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 04:56:51 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:56:51 localhost podman[300431]: 2025-11-27 09:56:51.124508953 +0000 UTC m=+0.201777984 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, container_name=openstack_network_exporter, release=1755695350, version=9.6, distribution-scope=public, io.openshift.tags=minimal rhel9, architecture=x86_64, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.buildah.version=1.33.7, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git) Nov 27 04:56:51 localhost podman[300431]: 2025-11-27 09:56:51.136907007 +0000 UTC m=+0.214176088 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, maintainer=Red Hat, Inc., build-date=2025-08-20T13:12:41, architecture=x86_64, managed_by=edpm_ansible, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.openshift.expose-services=, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm) Nov 27 04:56:51 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:56:51 localhost ceph-mon[291772]: Reconfig service osd.default_drive_group Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:51 localhost ceph-mon[291772]: from='mgr.26620 172.18.0.108:0/2380729673' entity='mgr.np0005537446.nfulyg' Nov 27 04:56:52 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e12 handle_command mon_command({"prefix": "mgr fail"} v 0) Nov 27 04:56:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='client.? 172.18.0.200:0/3108573532' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 e87: 6 total, 6 up, 6 in Nov 27 04:56:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:52.387+0000 7fae7966a640 -1 mgr handle_mgr_map I was active but no longer am Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr handle_mgr_map I was active but no longer am Nov 27 04:56:52 localhost nova_compute[284026]: 2025-11-27 09:56:52.441 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:52 localhost systemd[1]: session-68.scope: Deactivated successfully. Nov 27 04:56:52 localhost systemd[1]: session-68.scope: Consumed 7.427s CPU time. Nov 27 04:56:52 localhost systemd-logind[761]: Session 68 logged out. Waiting for processes to exit. Nov 27 04:56:52 localhost systemd-logind[761]: Removed session 68. Nov 27 04:56:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: ignoring --setuser ceph since I am not root Nov 27 04:56:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: ignoring --setgroup ceph since I am not root Nov 27 04:56:52 localhost ceph-mgr[290377]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-mgr, pid 2 Nov 27 04:56:52 localhost ceph-mgr[290377]: pidfile_write: ignore empty --pid-file Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr[py] Loading python module 'alerts' Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr[py] Module alerts has missing NOTIFY_TYPES member Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr[py] Loading python module 'balancer' Nov 27 04:56:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:52.666+0000 7f5572bf7140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr[py] Module balancer has missing NOTIFY_TYPES member Nov 27 04:56:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:52.733+0000 7f5572bf7140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member Nov 27 04:56:52 localhost ceph-mgr[290377]: mgr[py] Loading python module 'cephadm' Nov 27 04:56:52 localhost sshd[300491]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:56:52 localhost ceph-mon[291772]: from='client.? 172.18.0.200:0/3108573532' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: Activating manager daemon np0005537444.utkloe Nov 27 04:56:52 localhost ceph-mon[291772]: from='client.? ' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:56:52 localhost ceph-mon[291772]: Manager daemon np0005537444.utkloe is now available Nov 27 04:56:52 localhost ceph-mon[291772]: removing stray HostCache host record np0005537442.localdomain.devices.0 Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain.devices.0"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain.devices.0"}]': finished Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain.devices.0"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537442.localdomain.devices.0"}]': finished Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537444.utkloe/mirror_snapshot_schedule"} : dispatch Nov 27 04:56:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537444.utkloe/trash_purge_schedule"} : dispatch Nov 27 04:56:52 localhost systemd-logind[761]: New session 69 of user ceph-admin. Nov 27 04:56:52 localhost systemd[1]: Started Session 69 of User ceph-admin. Nov 27 04:56:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:56:53 localhost podman[300513]: 2025-11-27 09:56:53.143433005 +0000 UTC m=+0.096439328 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:56:53 localhost podman[300513]: 2025-11-27 09:56:53.157844892 +0000 UTC m=+0.110851265 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:56:53 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Loading python module 'crash' Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Module crash has missing NOTIFY_TYPES member Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Loading python module 'dashboard' Nov 27 04:56:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:53.358+0000 7f5572bf7140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Loading python module 'devicehealth' Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Module devicehealth has missing NOTIFY_TYPES member Nov 27 04:56:53 localhost ceph-mgr[290377]: mgr[py] Loading python module 'diskprediction_local' Nov 27 04:56:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:53.898+0000 7f5572bf7140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost podman[300629]: 2025-11-27 09:56:54.020039312 +0000 UTC m=+0.079055451 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, GIT_BRANCH=main, ceph=True, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., distribution-scope=public, RELEASE=main, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, version=7, build-date=2025-09-24T08:57:55, name=rhceph, GIT_CLEAN=True, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph) Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: from numpy import show_config as show_numpy_config Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:54.029+0000 7f5572bf7140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'influx' Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Module influx has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:54.086+0000 7f5572bf7140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'insights' Nov 27 04:56:54 localhost podman[300629]: 2025-11-27 09:56:54.132734365 +0000 UTC m=+0.191750504 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, CEPH_POINT_RELEASE=, architecture=x86_64, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, release=553, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, RELEASE=main, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'iostat' Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Module iostat has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'k8sevents' Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:54.202+0000 7f5572bf7140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'localpool' Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'mds_autoscaler' Nov 27 04:56:54 localhost nova_compute[284026]: 2025-11-27 09:56:54.702 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'mirroring' Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'nfs' Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Module nfs has missing NOTIFY_TYPES member Nov 27 04:56:54 localhost ceph-mgr[290377]: mgr[py] Loading python module 'orchestrator' Nov 27 04:56:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:54.968+0000 7f5572bf7140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module orchestrator has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'osd_perf_query' Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.116+0000 7f5572bf7140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'osd_support' Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.178+0000 7f5572bf7140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module osd_support has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'pg_autoscaler' Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.233+0000 7f5572bf7140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.312+0000 7f5572bf7140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'progress' Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module progress has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'prometheus' Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.375+0000 7f5572bf7140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mon[291772]: [27/Nov/2025:09:56:54] ENGINE Bus STARTING Nov 27 04:56:55 localhost ceph-mon[291772]: [27/Nov/2025:09:56:54] ENGINE Serving on https://172.18.0.106:7150 Nov 27 04:56:55 localhost ceph-mon[291772]: [27/Nov/2025:09:56:54] ENGINE Client ('172.18.0.106', 46466) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:56:55 localhost ceph-mon[291772]: [27/Nov/2025:09:56:54] ENGINE Serving on http://172.18.0.106:8765 Nov 27 04:56:55 localhost ceph-mon[291772]: [27/Nov/2025:09:56:54] ENGINE Bus STARTED Nov 27 04:56:55 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_DAEMON (was: 1 stray daemon(s) not managed by cephadm) Nov 27 04:56:55 localhost ceph-mon[291772]: Health check cleared: CEPHADM_STRAY_HOST (was: 1 stray host(s) with 1 daemon(s) not managed by cephadm) Nov 27 04:56:55 localhost ceph-mon[291772]: Health check cleared: MGR_MODULE_ERROR (was: Module 'cephadm' has failed: 'np0005537442.localdomain') Nov 27 04:56:55 localhost ceph-mon[291772]: Cluster is now healthy Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module prometheus has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.692+0000 7f5572bf7140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rbd_support' Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Module rbd_support has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:55.776+0000 7f5572bf7140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'restful' Nov 27 04:56:55 localhost openstack_network_exporter[244641]: ERROR 09:56:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:56:55 localhost openstack_network_exporter[244641]: ERROR 09:56:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:56:55 localhost openstack_network_exporter[244641]: ERROR 09:56:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:56:55 localhost openstack_network_exporter[244641]: ERROR 09:56:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:56:55 localhost openstack_network_exporter[244641]: Nov 27 04:56:55 localhost openstack_network_exporter[244641]: ERROR 09:56:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:56:55 localhost openstack_network_exporter[244641]: Nov 27 04:56:55 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rgw' Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Module rgw has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'rook' Nov 27 04:56:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:56.119+0000 7f5572bf7140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Module rook has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'selftest' Nov 27 04:56:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:56.581+0000 7f5572bf7140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Module selftest has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'snap_schedule' Nov 27 04:56:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:56.640+0000 7f5572bf7140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'stats' Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'status' Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Module status has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'telegraf' Nov 27 04:56:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:56.827+0000 7f5572bf7140 -1 mgr[py] Module status has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Module telegraf has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mgr[290377]: mgr[py] Loading python module 'telemetry' Nov 27 04:56:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:56.885+0000 7f5572bf7140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd/host:np0005537443", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:56:56 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:56:56 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:56:56 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:56:56 localhost ceph-mon[291772]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:56:56 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:56:56 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:56:56 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:56:56 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:56:56 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Module telemetry has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Loading python module 'test_orchestrator' Nov 27 04:56:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:57.014+0000 7f5572bf7140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Loading python module 'volumes' Nov 27 04:56:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:57.158+0000 7f5572bf7140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Module volumes has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Loading python module 'zabbix' Nov 27 04:56:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:57.347+0000 7f5572bf7140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: mgr[py] Module zabbix has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:56:57.407+0000 7f5572bf7140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member Nov 27 04:56:57 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x56195ce59600 mon_map magic: 0 from mon.1 v2:172.18.0.108:3300/0 Nov 27 04:56:57 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.106:6810/2385120390 Nov 27 04:56:57 localhost nova_compute[284026]: 2025-11-27 09:56:57.483 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:56:58 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:56:58 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:56:58 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:56:58 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:56:58 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.106:6810/2385120390 Nov 27 04:56:59 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:56:59 localhost nova_compute[284026]: 2025-11-27 09:56:59.726 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:00 localhost ceph-mon[291772]: Health check failed: 2 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:57:00 localhost ceph-mon[291772]: Health check failed: 2 stray host(s) with 2 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:57:00 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537443.fgpbcx (monmap changed)... Nov 27 04:57:00 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537443.fgpbcx", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:00 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537443.fgpbcx on np0005537443.localdomain Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:01 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:01 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:01 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:02 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:57:02 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:57:02 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:02 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:02 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:57:02 localhost nova_compute[284026]: 2025-11-27 09:57:02.518 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:57:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:57:03 localhost systemd[1]: tmp-crun.VExklr.mount: Deactivated successfully. Nov 27 04:57:03 localhost podman[301527]: 2025-11-27 09:57:03.01988023 +0000 UTC m=+0.104572596 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:57:03 localhost podman[301527]: 2025-11-27 09:57:03.054654284 +0000 UTC m=+0.139346590 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:57:03 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:57:03 localhost podman[301528]: 2025-11-27 09:57:03.073698685 +0000 UTC m=+0.157842076 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 04:57:03 localhost podman[301528]: 2025-11-27 09:57:03.081538754 +0000 UTC m=+0.165682085 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:57:03 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:57:03 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:57:03 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:57:03 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:04 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:04 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:04 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:04 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:04 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:04 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:57:04 localhost nova_compute[284026]: 2025-11-27 09:57:04.757 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:05 localhost ceph-mon[291772]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:57:05 localhost ceph-mon[291772]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:57:05 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:05 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:05 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:05 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:05 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:06 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:57:06 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:57:06 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:06 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:06 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:06 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:06 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:07 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:57:07 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:57:07 localhost ceph-mon[291772]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:57:07 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:07 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:57:07 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:07 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:07 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:07 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:57:07 localhost nova_compute[284026]: 2025-11-27 09:57:07.554 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:08 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:57:08 localhost ceph-mon[291772]: Reconfiguring osd.1 (monmap changed)... Nov 27 04:57:08 localhost ceph-mon[291772]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:57:08 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:08 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:08 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:08 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:08 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:57:08 localhost podman[242678]: time="2025-11-27T09:57:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:57:08 localhost podman[242678]: @ - - [27/Nov/2025:09:57:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:57:08 localhost podman[242678]: @ - - [27/Nov/2025:09:57:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19704 "" "Go-http-client/1.1" Nov 27 04:57:09 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:09 localhost ceph-mon[291772]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:57:09 localhost ceph-mon[291772]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:57:09 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:09 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:09 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:09 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:09 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:57:09 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:09 localhost nova_compute[284026]: 2025-11-27 09:57:09.794 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:10 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:57:10 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:10 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:10 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:57:10 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:10 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:57:11 localhost nova_compute[284026]: 2025-11-27 09:57:11.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:12 localhost ceph-mon[291772]: Reconfiguring mon.np0005537445 (monmap changed)... Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:12 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:12 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:12 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e12 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:57:12 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.107:0/3933504122' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:57:12 localhost nova_compute[284026]: 2025-11-27 09:57:12.587 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:12 localhost nova_compute[284026]: 2025-11-27 09:57:12.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:12 localhost nova_compute[284026]: 2025-11-27 09:57:12.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:12 localhost podman[301627]: Nov 27 04:57:12 localhost podman[301627]: 2025-11-27 09:57:12.925115538 +0000 UTC m=+0.082180245 container create 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., ceph=True, build-date=2025-09-24T08:57:55, RELEASE=main, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, version=7, com.redhat.component=rhceph-container, name=rhceph, io.buildah.version=1.33.12, GIT_CLEAN=True, release=553, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=) Nov 27 04:57:12 localhost systemd[1]: Started libpod-conmon-8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf.scope. Nov 27 04:57:12 localhost podman[301627]: 2025-11-27 09:57:12.889306777 +0000 UTC m=+0.046371574 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:12 localhost systemd[1]: Started libcrun container. Nov 27 04:57:13 localhost podman[301627]: 2025-11-27 09:57:13.010565101 +0000 UTC m=+0.167630008 container init 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, vcs-type=git, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., RELEASE=main, release=553, io.openshift.expose-services=, GIT_CLEAN=True, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, io.openshift.tags=rhceph ceph) Nov 27 04:57:13 localhost podman[301627]: 2025-11-27 09:57:13.021156244 +0000 UTC m=+0.178220951 container start 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , GIT_CLEAN=True, vendor=Red Hat, Inc., RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, GIT_BRANCH=main, version=7, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, architecture=x86_64, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, ceph=True, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, io.openshift.tags=rhceph ceph) Nov 27 04:57:13 localhost podman[301627]: 2025-11-27 09:57:13.021626718 +0000 UTC m=+0.178691465 container attach 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, io.buildah.version=1.33.12, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, version=7, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , name=rhceph, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., RELEASE=main, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:57:13 localhost angry_mestorf[301642]: 167 167 Nov 27 04:57:13 localhost systemd[1]: libpod-8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf.scope: Deactivated successfully. Nov 27 04:57:13 localhost podman[301627]: 2025-11-27 09:57:13.02881467 +0000 UTC m=+0.185879437 container died 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, architecture=x86_64, CEPH_POINT_RELEASE=, RELEASE=main, version=7, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , io.openshift.expose-services=, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.buildah.version=1.33.12, release=553, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:57:13 localhost podman[301647]: 2025-11-27 09:57:13.131177287 +0000 UTC m=+0.093341246 container remove 8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=angry_mestorf, GIT_CLEAN=True, name=rhceph, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, io.openshift.expose-services=, version=7, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:13 localhost systemd[1]: libpod-conmon-8cac520a398c3be8ec36ea9116f20e893170b36ad57feb1fd99b63992696e7cf.scope: Deactivated successfully. Nov 27 04:57:13 localhost ceph-mon[291772]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #28. Immutable memtables: 0. Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.421900) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 13] Flushing memtable with next log file: 28 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433421953, "job": 13, "event": "flush_started", "num_memtables": 1, "num_entries": 2801, "num_deletes": 255, "total_data_size": 7950371, "memory_usage": 8545824, "flush_reason": "Manual Compaction"} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 13] Level-0 flush table #29: started Nov 27 04:57:13 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:57:13 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:13 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:13 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433491317, "cf_name": "default", "job": 13, "event": "table_file_creation", "file_number": 29, "file_size": 4857012, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 17178, "largest_seqno": 19974, "table_properties": {"data_size": 4846149, "index_size": 6600, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3205, "raw_key_size": 28846, "raw_average_key_size": 22, "raw_value_size": 4821978, "raw_average_value_size": 3793, "num_data_blocks": 283, "num_entries": 1271, "num_filter_entries": 1271, "num_deletions": 253, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237341, "oldest_key_time": 1764237341, "file_creation_time": 1764237433, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 29, "seqno_to_time_mapping": "N/A"}} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 13] Flush lasted 69492 microseconds, and 10176 cpu microseconds. Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.491389) [db/flush_job.cc:967] [default] [JOB 13] Level-0 flush table #29: 4857012 bytes OK Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.491412) [db/memtable_list.cc:519] [default] Level-0 commit table #29 started Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.500770) [db/memtable_list.cc:722] [default] Level-0 commit table #29: memtable #1 done Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.500789) EVENT_LOG_v1 {"time_micros": 1764237433500783, "job": 13, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.500813) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 13] Try to delete WAL files size 7937013, prev total WAL file size 7937013, number of live WAL files 2. Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000025.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.501732) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003130373933' seq:72057594037927935, type:22 .. '7061786F73003131303435' seq:0, type:0; will stop at (end) Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 14] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 13 Base level 0, inputs: [29(4743KB)], [27(15MB)] Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433501791, "job": 14, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [29], "files_L6": [27], "score": -1, "input_data_size": 21319217, "oldest_snapshot_seqno": -1} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 14] Generated table #30: 10910 keys, 19145138 bytes, temperature: kUnknown Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433629477, "cf_name": "default", "job": 14, "event": "table_file_creation", "file_number": 30, "file_size": 19145138, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19081437, "index_size": 35312, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 27333, "raw_key_size": 292208, "raw_average_key_size": 26, "raw_value_size": 18893709, "raw_average_value_size": 1731, "num_data_blocks": 1349, "num_entries": 10910, "num_filter_entries": 10910, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237433, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 30, "seqno_to_time_mapping": "N/A"}} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.629907) [db/compaction/compaction_job.cc:1663] [default] [JOB 14] Compacted 1@0 + 1@6 files to L6 => 19145138 bytes Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.635026) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 166.8 rd, 149.8 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(4.6, 15.7 +0.0 blob) out(18.3 +0.0 blob), read-write-amplify(8.3) write-amplify(3.9) OK, records in: 11462, records dropped: 552 output_compression: NoCompression Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.635060) EVENT_LOG_v1 {"time_micros": 1764237433635046, "job": 14, "event": "compaction_finished", "compaction_time_micros": 127838, "compaction_time_cpu_micros": 23480, "output_level": 6, "num_output_files": 1, "total_output_size": 19145138, "num_input_records": 11462, "num_output_records": 10910, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000029.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433635783, "job": 14, "event": "table_file_deletion", "file_number": 29} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000027.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237433637991, "job": 14, "event": "table_file_deletion", "file_number": 27} Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.501635) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.638122) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.638130) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.638134) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.638137) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:13.638140) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:13 localhost nova_compute[284026]: 2025-11-27 09:57:13.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:13 localhost nova_compute[284026]: 2025-11-27 09:57:13.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:57:13 localhost systemd[1]: var-lib-containers-storage-overlay-e054405b716c1695ba56edd15e3613f1dbccbef60738fca76b38171689c70c27-merged.mount: Deactivated successfully. Nov 27 04:57:13 localhost podman[301717]: Nov 27 04:57:13 localhost podman[301717]: 2025-11-27 09:57:13.963397053 +0000 UTC m=+0.080322197 container create 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, name=rhceph, version=7, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, GIT_CLEAN=True, architecture=x86_64, maintainer=Guillaume Abrioux , ceph=True, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:14 localhost systemd[1]: Started libpod-conmon-8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b.scope. Nov 27 04:57:14 localhost systemd[1]: Started libcrun container. Nov 27 04:57:14 localhost podman[301717]: 2025-11-27 09:57:14.023009842 +0000 UTC m=+0.139935006 container init 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, GIT_BRANCH=main, CEPH_POINT_RELEASE=, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, release=553, name=rhceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, GIT_CLEAN=True, vcs-type=git, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:57:14 localhost podman[301717]: 2025-11-27 09:57:14.028822477 +0000 UTC m=+0.145747641 container start 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , version=7, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., ceph=True, com.redhat.component=rhceph-container, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:57:14 localhost podman[301717]: 2025-11-27 09:57:14.029166326 +0000 UTC m=+0.146091490 container attach 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, architecture=x86_64, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, io.openshift.expose-services=, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, release=553, name=rhceph, maintainer=Guillaume Abrioux , ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, version=7, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12) Nov 27 04:57:14 localhost romantic_bouman[301732]: 167 167 Nov 27 04:57:14 localhost systemd[1]: libpod-8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b.scope: Deactivated successfully. Nov 27 04:57:14 localhost podman[301717]: 2025-11-27 09:57:14.031762146 +0000 UTC m=+0.148687320 container died 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, CEPH_POINT_RELEASE=, vcs-type=git, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, release=553, architecture=x86_64, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, ceph=True) Nov 27 04:57:14 localhost podman[301717]: 2025-11-27 09:57:13.933833399 +0000 UTC m=+0.050758593 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:14 localhost podman[301737]: 2025-11-27 09:57:14.123434926 +0000 UTC m=+0.083937323 container remove 8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_bouman, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, io.openshift.expose-services=, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, com.redhat.component=rhceph-container, GIT_BRANCH=main, RELEASE=main, release=553, ceph=True, architecture=x86_64, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, name=rhceph, version=7, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:57:14 localhost systemd[1]: libpod-conmon-8bf8b3ad9e8ea6a786f2ea7216434211d620b3fdebf337b06548952fe2d2029b.scope: Deactivated successfully. Nov 27 04:57:14 localhost ceph-mon[291772]: mon.np0005537446@1(peon).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:14 localhost ceph-mon[291772]: Reconfiguring osd.0 (monmap changed)... Nov 27 04:57:14 localhost ceph-mon[291772]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:57:14 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:14 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:14 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:14 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:14 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch Nov 27 04:57:14 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e12 handle_command mon_command({"prefix": "mgr stat", "format": "json"} v 0) Nov 27 04:57:14 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.200:0/1321093153' entity='client.admin' cmd={"prefix": "mgr stat", "format": "json"} : dispatch Nov 27 04:57:14 localhost nova_compute[284026]: 2025-11-27 09:57:14.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:14 localhost nova_compute[284026]: 2025-11-27 09:57:14.797 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:14 localhost systemd[1]: var-lib-containers-storage-overlay-326135d43ebe7cec15e6145b3f311b2abf8ce79fe24fbc6429f8077e654190da-merged.mount: Deactivated successfully. Nov 27 04:57:14 localhost podman[301815]: Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:15.000097984 +0000 UTC m=+0.079379290 container create 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, build-date=2025-09-24T08:57:55, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, architecture=x86_64, vendor=Red Hat, Inc., io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:15 localhost systemd[1]: Started libpod-conmon-8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638.scope. Nov 27 04:57:15 localhost systemd[1]: Started libcrun container. Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:14.966913444 +0000 UTC m=+0.046194760 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:15.076665578 +0000 UTC m=+0.155946864 container init 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, name=rhceph, GIT_CLEAN=True, architecture=x86_64, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., version=7, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, distribution-scope=public, vcs-type=git) Nov 27 04:57:15 localhost blissful_lehmann[301831]: 167 167 Nov 27 04:57:15 localhost systemd[1]: libpod-8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638.scope: Deactivated successfully. Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:15.088460354 +0000 UTC m=+0.167741640 container start 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, architecture=x86_64, com.redhat.component=rhceph-container, RELEASE=main, version=7, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, release=553, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:15.088882126 +0000 UTC m=+0.168163442 container attach 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, name=rhceph, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, io.openshift.expose-services=, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, RELEASE=main, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:57:15 localhost podman[301815]: 2025-11-27 09:57:15.091498986 +0000 UTC m=+0.170780302 container died 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.openshift.expose-services=, version=7, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , RELEASE=main, vcs-type=git, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:15 localhost podman[301836]: 2025-11-27 09:57:15.244736787 +0000 UTC m=+0.143991524 container remove 8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=blissful_lehmann, maintainer=Guillaume Abrioux , io.openshift.expose-services=, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, name=rhceph, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, RELEASE=main, ceph=True, io.buildah.version=1.33.12) Nov 27 04:57:15 localhost systemd[1]: libpod-conmon-8b5bf3d11816e4dbbdb64831b2ae4512e1009c74927d1438b9a229c06dd80638.scope: Deactivated successfully. Nov 27 04:57:15 localhost ceph-mon[291772]: Reconfiguring osd.3 (monmap changed)... Nov 27 04:57:15 localhost ceph-mon[291772]: Reconfiguring daemon osd.3 on np0005537446.localdomain Nov 27 04:57:15 localhost nova_compute[284026]: 2025-11-27 09:57:15.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:15 localhost nova_compute[284026]: 2025-11-27 09:57:15.771 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:15 localhost systemd[1]: var-lib-containers-storage-overlay-fdf287a37d49bbf40ca8eafb16486dea8d8eb5da8826327604e00af025728ce0-merged.mount: Deactivated successfully. Nov 27 04:57:16 localhost podman[301914]: Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.134851026 +0000 UTC m=+0.082119754 container create 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, GIT_CLEAN=True, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, architecture=x86_64, name=rhceph, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, version=7, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, GIT_BRANCH=main, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:57:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:57:16 localhost systemd[1]: Started libpod-conmon-36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad.scope. Nov 27 04:57:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:57:16 localhost systemd[1]: Started libcrun container. Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.106432424 +0000 UTC m=+0.053701192 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.222301452 +0000 UTC m=+0.169570180 container init 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, release=553, GIT_CLEAN=True, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, vcs-type=git, name=rhceph, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, RELEASE=main, distribution-scope=public) Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.236564694 +0000 UTC m=+0.183833442 container start 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, io.buildah.version=1.33.12, ceph=True, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, io.openshift.expose-services=, version=7, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, distribution-scope=public, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, architecture=x86_64, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main) Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.236882173 +0000 UTC m=+0.184150891 container attach 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_BRANCH=main, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, name=rhceph, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:57:16 localhost determined_hertz[301930]: 167 167 Nov 27 04:57:16 localhost systemd[1]: libpod-36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad.scope: Deactivated successfully. Nov 27 04:57:16 localhost podman[301914]: 2025-11-27 09:57:16.241649641 +0000 UTC m=+0.188918379 container died 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, version=7, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, release=553, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, description=Red Hat Ceph Storage 7) Nov 27 04:57:16 localhost podman[301931]: 2025-11-27 09:57:16.328638285 +0000 UTC m=+0.136834522 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125) Nov 27 04:57:16 localhost podman[301931]: 2025-11-27 09:57:16.334881462 +0000 UTC m=+0.143077719 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:57:16 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:57:16 localhost podman[301929]: 2025-11-27 09:57:16.377169906 +0000 UTC m=+0.191687603 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=multipathd, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 04:57:16 localhost podman[301957]: 2025-11-27 09:57:16.450894055 +0000 UTC m=+0.197476429 container remove 36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_hertz, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , ceph=True, build-date=2025-09-24T08:57:55, name=rhceph, version=7, description=Red Hat Ceph Storage 7, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-type=git, distribution-scope=public, architecture=x86_64, GIT_BRANCH=main, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:57:16 localhost systemd[1]: libpod-conmon-36ade6442c8e2afbd9688110d23e30807b5b910a53006f3543e413a52c53e5ad.scope: Deactivated successfully. Nov 27 04:57:16 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:16 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:16 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:16 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:16 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537446.hkzfou (monmap changed)... Nov 27 04:57:16 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:16 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537446.hkzfou on np0005537446.localdomain Nov 27 04:57:16 localhost podman[301929]: 2025-11-27 09:57:16.469856004 +0000 UTC m=+0.284373701 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 04:57:16 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:57:16 localhost nova_compute[284026]: 2025-11-27 09:57:16.771 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:16 localhost systemd[1]: tmp-crun.zNqQXk.mount: Deactivated successfully. Nov 27 04:57:16 localhost systemd[1]: var-lib-containers-storage-overlay-54cc81141eadb3ccd4981a359b106ecfdc572f02e2b8b00d75c0f6e22dbca4fe-merged.mount: Deactivated successfully. Nov 27 04:57:17 localhost podman[302040]: Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.229129763 +0000 UTC m=+0.082000462 container create 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, architecture=x86_64, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , vcs-type=git, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, release=553, io.openshift.expose-services=, version=7, description=Red Hat Ceph Storage 7) Nov 27 04:57:17 localhost systemd[1]: Started libpod-conmon-6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689.scope. Nov 27 04:57:17 localhost systemd[1]: Started libcrun container. Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.195208142 +0000 UTC m=+0.048078861 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.301482603 +0000 UTC m=+0.154353312 container init 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, CEPH_POINT_RELEASE=, distribution-scope=public, description=Red Hat Ceph Storage 7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, io.openshift.expose-services=, vendor=Red Hat, Inc., vcs-type=git, name=rhceph, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, release=553, ceph=True, version=7, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.311128892 +0000 UTC m=+0.163999591 container start 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, version=7, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.buildah.version=1.33.12, GIT_BRANCH=main) Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.311638686 +0000 UTC m=+0.164509425 container attach 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, vcs-type=git, ceph=True, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, GIT_BRANCH=main, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, architecture=x86_64, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:57:17 localhost cranky_euclid[302055]: 167 167 Nov 27 04:57:17 localhost systemd[1]: libpod-6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689.scope: Deactivated successfully. Nov 27 04:57:17 localhost podman[302040]: 2025-11-27 09:57:17.315431967 +0000 UTC m=+0.168302706 container died 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, release=553, maintainer=Guillaume Abrioux , architecture=x86_64, version=7, CEPH_POINT_RELEASE=, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, name=rhceph, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:57:17 localhost podman[302060]: 2025-11-27 09:57:17.418847682 +0000 UTC m=+0.091456494 container remove 6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=cranky_euclid, vcs-type=git, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, maintainer=Guillaume Abrioux , version=7, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, name=rhceph, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, architecture=x86_64, CEPH_POINT_RELEASE=, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc.) Nov 27 04:57:17 localhost systemd[1]: libpod-conmon-6af431c75defd9fbb3c860f40edf709b193efcdd9f9f58fdaca159685df77689.scope: Deactivated successfully. Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:17 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537446.nfulyg (monmap changed)... Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:17 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537446.nfulyg on np0005537446.localdomain Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:17 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.635 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:17 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x56195ce58f20 mon_map magic: 0 from mon.1 v2:172.18.0.108:3300/0 Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@1(peon) e13 my rank is now 0 (was 1) Nov 27 04:57:17 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.108:3300/0 Nov 27 04:57:17 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.108:3300/0 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:57:17 localhost ceph-mon[291772]: paxos.0).electionLogic(60) init, last seen epoch 60 Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.763 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.763 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.764 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:57:17 localhost nova_compute[284026]: 2025-11-27 09:57:17.764 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e13 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445 in quorum (ranks 0,1,2) Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537444"} v 0) Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537444"} : dispatch Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537445"} v 0) Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537445"} : dispatch Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537446"} v 0) Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537446"} : dispatch Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : monmap epoch 13 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : last_changed 2025-11-27T09:57:17.749410+0000 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : created 2025-11-27T07:44:57.426149+0000 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : min_mon_release 18 (reef) Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : election_strategy: 1 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 0: [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] mon.np0005537446 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 1: [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] mon.np0005537444 Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 2: [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] mon.np0005537445 Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsmap cephfs:1 {0=mds.np0005537445.rkchqo=up:active} 2 up:standby Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : osdmap e87: 6 total, 6 up, 6 in Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : mgrmap e34: np0005537444.utkloe(active, since 25s), standbys: np0005537445.wmiblp, np0005537443.fgpbcx, np0005537446.nfulyg, np0005537441.xbypnw Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:17 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:17 localhost ceph-mon[291772]: Reconfiguring mon.np0005537446 (monmap changed)... Nov 27 04:57:17 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:57:17 localhost ceph-mon[291772]: Remove daemons mon.np0005537443 Nov 27 04:57:17 localhost ceph-mon[291772]: Safe to remove mon.np0005537443: new quorum should be ['np0005537446', 'np0005537444', 'np0005537445'] (from ['np0005537446', 'np0005537444', 'np0005537445']) Nov 27 04:57:17 localhost ceph-mon[291772]: Removing monitor np0005537443 from monmap... Nov 27 04:57:17 localhost ceph-mon[291772]: Removing daemon mon.np0005537443 from np0005537443.localdomain -- ports [] Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445 in quorum (ranks 0,1,2) Nov 27 04:57:17 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:57:17 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:17 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:17 localhost ceph-mon[291772]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:17 localhost systemd[1]: var-lib-containers-storage-overlay-ca9b317d3ef8a0950cabc7c79f827643abee91a51523c4fd4c17a6c650abef66-merged.mount: Deactivated successfully. Nov 27 04:57:17 localhost ceph-mgr[290377]: --2- 172.18.0.108:0/2980308931 >> [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] conn(0x56195cd8f400 0x56195cdff080 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=0 rev1=1 crypto rx=0 tx=0 comp rx=0 tx=0).send_auth_request get_initial_auth_request returned -2 Nov 27 04:57:17 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.104:3300/0 Nov 27 04:57:17 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.104:3300/0 Nov 27 04:57:17 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x561963d9a000 mon_map magic: 0 from mon.0 v2:172.18.0.108:3300/0 Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.219 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.454s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:57:18 localhost podman[302150]: Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.235986843 +0000 UTC m=+0.090581740 container create 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, CEPH_POINT_RELEASE=, vcs-type=git, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, ceph=True, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, architecture=x86_64, GIT_CLEAN=True, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, io.openshift.tags=rhceph ceph) Nov 27 04:57:18 localhost systemd[1]: Started libpod-conmon-781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89.scope. Nov 27 04:57:18 localhost systemd[1]: Started libcrun container. Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.289 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.290 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.301662905 +0000 UTC m=+0.156257822 container init 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, architecture=x86_64, name=rhceph, release=553, io.openshift.expose-services=, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.openshift.tags=rhceph ceph, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., RELEASE=main, io.buildah.version=1.33.12) Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.201908029 +0000 UTC m=+0.056502936 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.3145036 +0000 UTC m=+0.169098487 container start 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, name=rhceph, GIT_BRANCH=main, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, CEPH_POINT_RELEASE=) Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.315004543 +0000 UTC m=+0.169599490 container attach 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, release=553, GIT_BRANCH=main, io.openshift.expose-services=, version=7, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, ceph=True, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., distribution-scope=public, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, name=rhceph, architecture=x86_64) Nov 27 04:57:18 localhost romantic_nash[302168]: 167 167 Nov 27 04:57:18 localhost systemd[1]: libpod-781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89.scope: Deactivated successfully. Nov 27 04:57:18 localhost podman[302150]: 2025-11-27 09:57:18.321069866 +0000 UTC m=+0.175664803 container died 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., version=7, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, build-date=2025-09-24T08:57:55, GIT_CLEAN=True, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, architecture=x86_64, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553) Nov 27 04:57:18 localhost podman[302173]: 2025-11-27 09:57:18.464505754 +0000 UTC m=+0.122850707 container remove 781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=romantic_nash, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, ceph=True, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, com.redhat.component=rhceph-container, io.openshift.expose-services=, release=553, vendor=Red Hat, Inc., name=rhceph, vcs-type=git, version=7, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:57:18 localhost systemd[1]: libpod-conmon-781bda6ffeeff8cb448ae1c24aa4245347da89887f7bc74101c730032dfa6b89.scope: Deactivated successfully. Nov 27 04:57:18 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.551 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.553 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11421MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.553 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.554 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:57:18 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:18 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:57:18 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:18 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:18 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:18 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 04:57:18 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.865 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.865 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.866 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:57:18 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:18 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:18 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:18 localhost systemd[1]: tmp-crun.2mXTFH.mount: Deactivated successfully. Nov 27 04:57:18 localhost systemd[1]: var-lib-containers-storage-overlay-df47fc83e91b129ab7c2a7e75c453870567880ab288f2d2bc8de5d7a703104a1-merged.mount: Deactivated successfully. Nov 27 04:57:18 localhost nova_compute[284026]: 2025-11-27 09:57:18.969 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.049 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.050 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.081 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.125 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.171 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:57:19 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:19 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:57:19 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1017274466' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.655 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.484s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.660 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.700 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.704 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.704 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.150s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.705 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.706 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.726 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.726 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.726 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.747 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:19 localhost nova_compute[284026]: 2025-11-27 09:57:19.846 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:19 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:57:19 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:19 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:20 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "auth get", "entity": "mon."} v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "config get", "who": "mon", "key": "public_network"} v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch Nov 27 04:57:20 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:20 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.764 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.765 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.765 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.905 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.905 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.905 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:57:20 localhost nova_compute[284026]: 2025-11-27 09:57:20.906 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:57:21 localhost nova_compute[284026]: 2025-11-27 09:57:21.268 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:57:21 localhost nova_compute[284026]: 2025-11-27 09:57:21.285 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:57:21 localhost nova_compute[284026]: 2025-11-27 09:57:21.286 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:57:21 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:21 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:57:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:57:22 localhost podman[302531]: 2025-11-27 09:57:22.092568575 +0000 UTC m=+0.095092122 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 04:57:22 localhost podman[302531]: 2025-11-27 09:57:22.104506545 +0000 UTC m=+0.107030122 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 04:57:22 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:57:22 localhost systemd[1]: tmp-crun.XJZnts.mount: Deactivated successfully. Nov 27 04:57:22 localhost podman[302532]: 2025-11-27 09:57:22.209331207 +0000 UTC m=+0.212044879 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, build-date=2025-08-20T13:12:41, maintainer=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter, vendor=Red Hat, Inc., name=ubi9-minimal, distribution-scope=public, io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.buildah.version=1.33.7, architecture=x86_64, com.redhat.component=ubi9-minimal-container, vcs-type=git, managed_by=edpm_ansible) Nov 27 04:57:22 localhost podman[302532]: 2025-11-27 09:57:22.22847677 +0000 UTC m=+0.231190402 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, config_id=edpm, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, distribution-scope=public, name=ubi9-minimal, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., vendor=Red Hat, Inc., release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, architecture=x86_64, vcs-type=git, container_name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible) Nov 27 04:57:22 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:57:22 localhost nova_compute[284026]: 2025-11-27 09:57:22.669 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:22 localhost ceph-mon[291772]: Deploying daemon mon.np0005537443 on np0005537443.localdomain Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:23 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 adding peer [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] to list of hints Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 adding peer [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] to list of hints Nov 27 04:57:23 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mon}] v 0) Nov 27 04:57:23 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mon}] v 0) Nov 27 04:57:23 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:23 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 04:57:23 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 04:57:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:57:24 localhost podman[302588]: 2025-11-27 09:57:23.999920553 +0000 UTC m=+0.099040728 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:57:24 localhost podman[302588]: 2025-11-27 09:57:24.039997048 +0000 UTC m=+0.139117213 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:24 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 adding peer [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] to list of hints Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(leader).monmap v13 adding/updating np0005537443 at [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] to monitor cluster Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e13 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:24 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x56195ce58f20 mon_map magic: 0 from mon.0 v2:172.18.0.108:3300/0 Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(probing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(probing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537444"} v 0) Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537444"} : dispatch Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(probing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537445"} v 0) Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537445"} : dispatch Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:57:24 localhost ceph-mon[291772]: paxos.0).electionLogic(62) init, last seen epoch 62 Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:24 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537446"} v 0) Nov 27 04:57:24 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537446"} : dispatch Nov 27 04:57:24 localhost nova_compute[284026]: 2025-11-27 09:57:24.891 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:25 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_auth_request failed to assign global_id Nov 27 04:57:25 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:25 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:25 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_auth_request failed to assign global_id Nov 27 04:57:25 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:25 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_auth_request failed to assign global_id Nov 27 04:57:25 localhost openstack_network_exporter[244641]: ERROR 09:57:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:57:25 localhost openstack_network_exporter[244641]: ERROR 09:57:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:57:25 localhost openstack_network_exporter[244641]: ERROR 09:57:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:57:25 localhost openstack_network_exporter[244641]: ERROR 09:57:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:57:25 localhost openstack_network_exporter[244641]: Nov 27 04:57:25 localhost openstack_network_exporter[244641]: ERROR 09:57:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:57:25 localhost openstack_network_exporter[244641]: Nov 27 04:57:26 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:26 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:26 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_auth_request failed to assign global_id Nov 27 04:57:27 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:27 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.461 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.480 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.481 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.482 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.508 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.026s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:57:27 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 04:57:27 localhost nova_compute[284026]: 2025-11-27 09:57:27.706 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:28 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:28 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:29 localhost ceph-mon[291772]: paxos.0).electionLogic(63) init, last seen epoch 63, mid-election, bumping Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e14 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445,np0005537443 in quorum (ranks 0,1,2,3) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : monmap epoch 14 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : last_changed 2025-11-27T09:57:24.063631+0000 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : created 2025-11-27T07:44:57.426149+0000 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : min_mon_release 18 (reef) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : election_strategy: 1 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 0: [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] mon.np0005537446 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 1: [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] mon.np0005537444 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 2: [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] mon.np0005537445 Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 3: [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] mon.np0005537443 Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsmap cephfs:1 {0=mds.np0005537445.rkchqo=up:active} 2 up:standby Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : osdmap e87: 6 total, 6 up, 6 in Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : mgrmap e34: np0005537444.utkloe(active, since 36s), standbys: np0005537445.wmiblp, np0005537443.fgpbcx, np0005537446.nfulyg, np0005537441.xbypnw Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:29 localhost ceph-mon[291772]: Removed label mon from host np0005537443.localdomain Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537443 calling monitor election Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445,np0005537443 in quorum (ranks 0,1,2,3) Nov 27 04:57:29 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:29 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:29 localhost ceph-mon[291772]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:29 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:29 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:29 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:29 localhost nova_compute[284026]: 2025-11-27 09:57:29.895 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537443"} v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537443"} : dispatch Nov 27 04:57:30 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:30 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:30 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:30 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:30 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:30 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: Removed label mgr from host np0005537443.localdomain Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:30 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 04:57:30 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: Updating np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:31 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:31 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:31 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:31 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:32 localhost ceph-mon[291772]: Removed label _admin from host np0005537443.localdomain Nov 27 04:57:32 localhost ceph-mon[291772]: Removing daemon mgr.np0005537443.fgpbcx from np0005537443.localdomain -- ports [8765] Nov 27 04:57:32 localhost nova_compute[284026]: 2025-11-27 09:57:32.709 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "auth rm", "entity": "mgr.np0005537443.fgpbcx"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth rm", "entity": "mgr.np0005537443.fgpbcx"} : dispatch Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd='[{"prefix": "auth rm", "entity": "mgr.np0005537443.fgpbcx"}]': finished Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mgr}] v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mgr}] v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "mon ok-to-stop", "ids": ["np0005537443"]} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon ok-to-stop", "ids": ["np0005537443"]} : dispatch Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "quorum_status"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "quorum_status"} : dispatch Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e14 handle_command mon_command({"prefix": "mon rm", "name": "np0005537443"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon rm", "name": "np0005537443"} : dispatch Nov 27 04:57:33 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x561963d90000 mon_map magic: 0 from mon.0 v2:172.18.0.108:3300/0 Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(probing) e15 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537444"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537444"} : dispatch Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:57:33 localhost ceph-mon[291772]: paxos.0).electionLogic(66) init, last seen epoch 66 Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537445"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537445"} : dispatch Nov 27 04:57:33 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537446"} v 0) Nov 27 04:57:33 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon metadata", "id": "np0005537446"} : dispatch Nov 27 04:57:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:57:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:57:33 localhost podman[302934]: 2025-11-27 09:57:33.991335853 +0000 UTC m=+0.078528058 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:57:34 localhost podman[302934]: 2025-11-27 09:57:34.005062901 +0000 UTC m=+0.092255136 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:57:34 localhost podman[302933]: 2025-11-27 09:57:34.04827622 +0000 UTC m=+0.137889730 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller) Nov 27 04:57:34 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:57:34 localhost podman[302933]: 2025-11-27 09:57:34.11646037 +0000 UTC m=+0.206073880 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.vendor=CentOS) Nov 27 04:57:34 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:57:34 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 04:57:34 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mon}] v 0) Nov 27 04:57:34 localhost nova_compute[284026]: 2025-11-27 09:57:34.923 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:37 localhost nova_compute[284026]: 2025-11-27 09:57:37.749 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 is new leader, mons np0005537446,np0005537445 in quorum (ranks 0,2) Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:57:38 localhost ceph-mon[291772]: paxos.0).electionLogic(69) init, last seen epoch 69, mid-election, bumping Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446@0(electing) e15 handle_timecheck drop unexpected msg Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [INF] : mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445 in quorum (ranks 0,1,2) Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : monmap epoch 15 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : last_changed 2025-11-27T09:57:33.096302+0000 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : created 2025-11-27T07:44:57.426149+0000 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : min_mon_release 18 (reef) Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : election_strategy: 1 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 0: [v2:172.18.0.108:3300/0,v1:172.18.0.108:6789/0] mon.np0005537446 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 1: [v2:172.18.0.103:3300/0,v1:172.18.0.103:6789/0] mon.np0005537444 Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : 2: [v2:172.18.0.104:3300/0,v1:172.18.0.104:6789/0] mon.np0005537445 Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : fsmap cephfs:1 {0=mds.np0005537445.rkchqo=up:active} 2 up:standby Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : osdmap e87: 6 total, 6 up, 6 in Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [DBG] : mgrmap e34: np0005537444.utkloe(active, since 45s), standbys: np0005537445.wmiblp, np0005537443.fgpbcx, np0005537446.nfulyg, np0005537441.xbypnw Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(cluster) log [WRN] : stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mon}] v 0) Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 04:57:38 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 04:57:38 localhost ceph-mon[291772]: Removing key for mgr.np0005537443.fgpbcx Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537445 calling monitor election Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:57:38 localhost ceph-mon[291772]: Safe to remove mon.np0005537443: new quorum should be ['np0005537446', 'np0005537444', 'np0005537445'] (from ['np0005537446', 'np0005537444', 'np0005537445']) Nov 27 04:57:38 localhost ceph-mon[291772]: Removing monitor np0005537443 from monmap... Nov 27 04:57:38 localhost ceph-mon[291772]: Removing daemon mon.np0005537443 from np0005537443.localdomain -- ports [] Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537444 calling monitor election Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446 is new leader, mons np0005537446,np0005537445 in quorum (ranks 0,2) Nov 27 04:57:38 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:38 localhost ceph-mon[291772]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446 calling monitor election Nov 27 04:57:38 localhost ceph-mon[291772]: mon.np0005537446 is new leader, mons np0005537446,np0005537444,np0005537445 in quorum (ranks 0,1,2) Nov 27 04:57:38 localhost ceph-mon[291772]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:57:38 localhost ceph-mon[291772]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:57:38 localhost ceph-mon[291772]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:57:38 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:38 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:38 localhost podman[242678]: time="2025-11-27T09:57:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:57:38 localhost podman[242678]: @ - - [27/Nov/2025:09:57:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:57:38 localhost podman[242678]: @ - - [27/Nov/2025:09:57:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19713 "" "Go-http-client/1.1" Nov 27 04:57:39 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:39 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:39 localhost nova_compute[284026]: 2025-11-27 09:57:39.926 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:39 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:39 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:39 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:39 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:39 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:39 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:39 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 04:57:39 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:40 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:40 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:40 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:40 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:40 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:40 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:40 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:57:40 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:40 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: Removing np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:41 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:41 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:41 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:57:41 localhost ceph-mon[291772]: Removing np0005537443.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:57:41 localhost ceph-mon[291772]: Removing np0005537443.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:57:41 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:41 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:41 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:42 localhost ceph-mon[291772]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:42 localhost ceph-mon[291772]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537443.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain.devices.0}] v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537443.localdomain}] v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:42 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:42 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:42 localhost nova_compute[284026]: 2025-11-27 09:57:42.791 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:43 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 04:57:43 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ceph-mon[291772]: Reconfiguring crash.np0005537443 (monmap changed)... Nov 27 04:57:43 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537443 on np0005537443.localdomain Nov 27 04:57:43 localhost ceph-mon[291772]: Added label _no_schedule to host np0005537443.localdomain Nov 27 04:57:43 localhost ceph-mon[291772]: Added label SpecialHostLabels.DRAIN_CONF_KEYRING to host np0005537443.localdomain Nov 27 04:57:43 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:43 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:57:43.567 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:57:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:57:43.567 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:57:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:57:43.568 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:57:43 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:43 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:43 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:43 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "osd.2"} v 0) Nov 27 04:57:43 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:57:43 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:43 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.138 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.140 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.153 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.154 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '526114c9-0382-40e6-a269-5e7c508cd60f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.140560', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '850221f4-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': 'e537b8059b70a9b1ed21d1924a484567404e962ad88fbd69182fa5370a7ddac7'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.140560', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '8502476a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': '94ce27fa833388c156bd6ab4c7be5bda0bded5e5f24475a18b340be642c35554'}]}, 'timestamp': '2025-11-27 09:57:44.155483', '_unique_id': 'be26ddd06db44be4a3eed2d5ab2ce0e6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.158 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.160 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.165 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '063d6add-2b5f-4931-ab2f-434253b49eff', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.160437', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '8503ff1a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': 'b925f95a82ff2a28911f09f08cbf6d56962e0ce092e69f357d4068d8dc9443d4'}]}, 'timestamp': '2025-11-27 09:57:44.166876', '_unique_id': '9ef9b7773bf74f8a84a0a0b12e10f344'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.170 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.170 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c4ee0a0a-1f62-4726-97be-a9717b5feb2d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.170878', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '8504bc66-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '367cdcebef24986e7aa1445438c9f37a3b738ef76dc6ad127490dfb83bea18b8'}]}, 'timestamp': '2025-11-27 09:57:44.171590', '_unique_id': '69ab0ed880814c37b607a78f1a9dd88a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.175 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.175 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8594195b-05e5-44fd-8053-4386d4b39a3d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.175224', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '85056814-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '5c2ad0a843b7ebd9aae0b129f17d2d10a65a66f75ddcf43369f18216fc30ce47'}]}, 'timestamp': '2025-11-27 09:57:44.175953', '_unique_id': '6bbd98bfa29847d38cbd125cba3114fc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.178 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.179 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.179 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.179 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '88b43d41-7094-4382-a154-cf97302f9bad', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.179741', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '85061782-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '1e2b687b12455883877523f8d42433b048793de44d7560353b598656b717de3d'}]}, 'timestamp': '2025-11-27 09:57:44.180461', '_unique_id': 'eb351d9239ce439f965b3b34f560c2e6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.181 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.183 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.183 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.199 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 13620000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bfe2e1a9-4651-444d-97b8-9ef6a814683f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 13620000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:57:44.183843', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '85091860-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.386996416, 'message_signature': '0378612a0bf2fc9aaf9f8d19fd5d8c8a307eff26b2bec9dcb91b2253ea61e486'}]}, 'timestamp': '2025-11-27 09:57:44.200181', '_unique_id': 'f03153adeade4e33bab860d6c4e4e312'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.203 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.204 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'df7b30e5-6102-41e7-a1c9-f8f22f545a6b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.203994', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '8509ca30-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '7b3a71c14fdf140b40186067859fb5a1637ca75cbfba96dac8f3c566a4154d75'}]}, 'timestamp': '2025-11-27 09:57:44.204726', '_unique_id': '51f4d0a6bfe94e18aa6811e78d8b7839'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.207 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.207 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd1d2c346-b67e-466a-97c0-42a7fb799e05', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.207895', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '850a5d60-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '8d4f51b1661bd559b4ad2d977887007c7ec961af98af37c2d79034082abc69a5'}]}, 'timestamp': '2025-11-27 09:57:44.208324', '_unique_id': 'be01e6db00584df193304529d07fdbfa'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.209 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.210 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.210 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9182f9c2-bb8b-4157-b9dc-50eb21818ba5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:57:44.210258', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '850ab832-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.386996416, 'message_signature': '7d0d7f50743a1e51ac4d33ee22922f35ea27845501f60a1dff97b61f51c93e1f'}]}, 'timestamp': '2025-11-27 09:57:44.210658', '_unique_id': '1b5b904d04594520a864a5084de47b40'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.212 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.243 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ff1c6528-9dca-4dcd-93c1-c04eb6c55ff4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.212645', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '850fdf2e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'c685e1993f7b3d98453614ad4dd36a445dba2295ebeff5421fe6c3925d9c93a9'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.212645', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '850fef78-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'df3fae26aefb7d4727277d402b464b8d7db1754cc46c96343815d04f34ce544b'}]}, 'timestamp': '2025-11-27 09:57:44.244803', '_unique_id': '42ecbf4684af4bc68d0561315645a283'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.246 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.246 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '52b1a2c2-347f-45c0-8c19-7cb527a1eb41', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.246875', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '85104fe0-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '27ba734ae57f929f7ee2db762e639e66adc90edbc8c92c2ef43c6806133123e3'}]}, 'timestamp': '2025-11-27 09:57:44.247249', '_unique_id': 'ba6850d223bd4167af448ca3869934b5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8d9be22e-9179-4108-97e3-db7d2049db11', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.248753', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '85109720-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '3a3856ce4886d4fca95404e148181c80c7251abedb130d4916bdb8fcc730e734'}]}, 'timestamp': '2025-11-27 09:57:44.249068', '_unique_id': '4cbce46bef4c41b1a69e3ea2397cd085'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.250 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4db725dd-f20b-4903-bc8c-b42b9d2f9ae4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.250589', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '8510def6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'cd95bf4d382f65273661e8347e185b78af950ea71c4f736a675591524fab15b0'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.250589', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '8510e9b4-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': '075f0039e2e7019e5ef6d32c43761fa00b7108a0ee1cdb50e603105bc3d3a9f7'}]}, 'timestamp': '2025-11-27 09:57:44.251182', '_unique_id': 'dcfff79bc9ce4f4881b17fba8a8abb64'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.252 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e2725f03-5858-4943-8b33-714ba39d0beb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.252639', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '85112e92-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'f77f71291619590610283f34acddc00beef9d4e08a7442bed48d549db6ad420e'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.252639', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '85113900-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': '331583f888ed1f05f2b449c49099bd6774f23171b443cc2afd979ec8b89abab4'}]}, 'timestamp': '2025-11-27 09:57:44.253188', '_unique_id': 'e4b820386ba24a99ba47c7a5de938809'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.254 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.254 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ccb1521e-2e63-4da9-a74f-9e8d1e871220', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.254687', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '85117f00-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'dd6d089c51751afd818fda47a55b17ea0d4c5ea8ad96c50ff1d36c8c3ec3a6bf'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.254687', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '85118c34-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'cb81da44a2dd2e4f0e945b445323a9367703b3a5e623c9bc450c0e108a3399db'}]}, 'timestamp': '2025-11-27 09:57:44.255322', '_unique_id': 'f48837183fe6435aaf0515b33624fdf7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.257 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a03e1de9-63d0-449a-a371-858f36d7ea2e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.257139', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '8511e030-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': '1a53f060a4472eda1f3ca43980bf1d98e64fcfba9d9f72bc8318fbc9c13f1eb2'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.257139', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '8511ed5a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': 'a889b69af518a531ee9b7761dbe5669638b73d5155ce447a3d9a077f9700dfca'}]}, 'timestamp': '2025-11-27 09:57:44.257810', '_unique_id': '287aae400a2f46e9b1ad6ca3f513b47f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.260 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.260 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.260 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd60d3d89-f81d-478d-a434-82c5d6ce36d2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.260299', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '85125ac4-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': '5a98648899670e2cc88fffdb83cf5f88094f47b9e0a7b09b395bcb0a0d6bb56c'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.260299', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '851266ea-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.328653231, 'message_signature': '5d71e92028508ff5ca3f2157ca6c5334ddde80c41b0a649d69c1f7e1cfe41ca9'}]}, 'timestamp': '2025-11-27 09:57:44.260921', '_unique_id': 'a1f8c4c09f9b42a9a6ac149a8a63f08c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.262 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.262 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.262 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1d92eddf-e1e1-43ee-9f0a-af4e95325379', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.262495', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '8512b0be-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'ff0957259a2aa8b9da165059141f064484c07d9ed18a36b8ec42d052433965ae'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.262495', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '8512bbae-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': '0c2b95b251bbd11b158198e2988670be223cc8685669c5596c9c43420989f1ea'}]}, 'timestamp': '2025-11-27 09:57:44.263131', '_unique_id': 'a318a1031d824042b63a6e2f0c03add8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.264 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '067d62a0-a8e1-4852-ba53-e3fa3ab2cf06', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.264804', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '85130a1e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': '267049c327f1f0d5f80d3ccc7f7fc04e275362f499d79c06f71ad6bd813cbfda'}]}, 'timestamp': '2025-11-27 09:57:44.265130', '_unique_id': '3f58b92f11c64d169fa427dfaa1957f8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.266 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.266 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.267 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '651f5dd6-9182-4ee6-9f81-17b061067310', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:57:44.266754', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '851357f8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': 'c878032ab9bba26d216517e8381d03e681105a767f3fd9eb5e79b48ef9e26c54'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:57:44.266754', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '8513687e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.400722954, 'message_signature': '0490be4f6726db9fdb46f856fcaa8193c9a24e04b26a0a2dd5649e596d79348f'}]}, 'timestamp': '2025-11-27 09:57:44.267608', '_unique_id': '7dedc961050f454cb45fa0f95b5a2e40'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.268 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.269 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.269 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '38611fdf-b5d6-4a9b-af34-41f3c61b8d69', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:57:44.269657', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '8513ca8a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11694.348609416, 'message_signature': 'e63e17c69c52d83b5c6535151870a7f530b38e64f7e76e7914855a0f5b7d99ac'}]}, 'timestamp': '2025-11-27 09:57:44.270118', '_unique_id': '1803f909e32243b9b44caa4563bb1146'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:57:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:57:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:44 localhost ceph-mon[291772]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:57:44 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:57:44 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:44 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:44 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "osd.5"} v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/inventory}] v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:44 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain"} v 0) Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain"} : dispatch Nov 27 04:57:44 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain"}]': finished Nov 27 04:57:44 localhost nova_compute[284026]: 2025-11-27 09:57:44.929 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:45 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:45 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:45 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} v 0) Nov 27 04:57:45 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:45 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:45 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:45 localhost ceph-mon[291772]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:57:45 localhost ceph-mon[291772]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain"} : dispatch Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain"}]': finished Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:45 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:46 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:46 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:46 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:46 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:46 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} v 0) Nov 27 04:57:46 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:46 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "mgr services"} v 0) Nov 27 04:57:46 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mgr services"} : dispatch Nov 27 04:57:46 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:46 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:57:46 localhost ceph-mon[291772]: Removed host np0005537443.localdomain Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:57:46 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:46 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:46 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:57:46 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:57:46 localhost sshd[303337]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:57:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:57:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:57:46 localhost systemd-logind[761]: New session 70 of user tripleo-admin. Nov 27 04:57:46 localhost systemd[1]: Created slice User Slice of UID 1003. Nov 27 04:57:47 localhost systemd[1]: Starting User Runtime Directory /run/user/1003... Nov 27 04:57:47 localhost systemd[1]: Finished User Runtime Directory /run/user/1003. Nov 27 04:57:47 localhost podman[303339]: 2025-11-27 09:57:47.034008599 +0000 UTC m=+0.122381665 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 04:57:47 localhost systemd[1]: Starting User Manager for UID 1003... Nov 27 04:57:47 localhost podman[303340]: 2025-11-27 09:57:47.065603526 +0000 UTC m=+0.149478261 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, config_id=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:57:47 localhost podman[303340]: 2025-11-27 09:57:47.083984109 +0000 UTC m=+0.167858864 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, managed_by=edpm_ansible) Nov 27 04:57:47 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:57:47 localhost podman[303339]: 2025-11-27 09:57:47.165899977 +0000 UTC m=+0.254273003 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:57:47 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:57:47 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:47 localhost systemd[303371]: Queued start job for default target Main User Target. Nov 27 04:57:47 localhost systemd[303371]: Created slice User Application Slice. Nov 27 04:57:47 localhost systemd[303371]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 04:57:47 localhost systemd[303371]: Started Daily Cleanup of User's Temporary Directories. Nov 27 04:57:47 localhost systemd[303371]: Reached target Paths. Nov 27 04:57:47 localhost systemd[303371]: Reached target Timers. Nov 27 04:57:47 localhost systemd[303371]: Starting D-Bus User Message Bus Socket... Nov 27 04:57:47 localhost systemd[303371]: Starting Create User's Volatile Files and Directories... Nov 27 04:57:47 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:47 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:47 localhost systemd[303371]: Listening on D-Bus User Message Bus Socket. Nov 27 04:57:47 localhost systemd[303371]: Reached target Sockets. Nov 27 04:57:47 localhost systemd[303371]: Finished Create User's Volatile Files and Directories. Nov 27 04:57:47 localhost systemd[303371]: Reached target Basic System. Nov 27 04:57:47 localhost systemd[303371]: Reached target Main User Target. Nov 27 04:57:47 localhost systemd[303371]: Startup finished in 200ms. Nov 27 04:57:47 localhost systemd[1]: Started User Manager for UID 1003. Nov 27 04:57:47 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:47 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "mon."} v 0) Nov 27 04:57:47 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:47 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config get", "who": "mon", "key": "public_network"} v 0) Nov 27 04:57:47 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch Nov 27 04:57:47 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:47 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:47 localhost systemd[1]: Started Session 70 of User tripleo-admin. Nov 27 04:57:47 localhost nova_compute[284026]: 2025-11-27 09:57:47.794 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:48 localhost python3[303522]: ansible-ansible.builtin.lineinfile Invoked with dest=/etc/os-net-config/tripleo_config.yaml insertafter=172.18.0 line= - ip_netmask: 172.18.0.105/24 backup=True path=/etc/os-net-config/tripleo_config.yaml state=present backrefs=False create=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Nov 27 04:57:48 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:57:48 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:57:48 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} v 0) Nov 27 04:57:48 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:48 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:48 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:48 localhost ceph-mon[291772]: Reconfiguring mon.np0005537444 (monmap changed)... Nov 27 04:57:48 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:48 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:48 localhost python3[303668]: ansible-ansible.legacy.command Invoked with _raw_params=ip a add 172.18.0.105/24 dev vlan21 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:57:49 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:49 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:49 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:49 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:49 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "osd.1"} v 0) Nov 27 04:57:49 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:57:49 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:49 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:49 localhost ceph-mon[291772]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:57:49 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:57:49 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:49 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:49 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:57:49 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #31. Immutable memtables: 0. Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.346003) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:856] [default] [JOB 15] Flushing memtable with next log file: 31 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469346111, "job": 15, "event": "flush_started", "num_memtables": 1, "num_entries": 1521, "num_deletes": 263, "total_data_size": 2575747, "memory_usage": 2621088, "flush_reason": "Manual Compaction"} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:885] [default] [JOB 15] Level-0 flush table #32: started Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469368193, "cf_name": "default", "job": 15, "event": "table_file_creation", "file_number": 32, "file_size": 1830127, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 19979, "largest_seqno": 21495, "table_properties": {"data_size": 1823677, "index_size": 3279, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2117, "raw_key_size": 17973, "raw_average_key_size": 21, "raw_value_size": 1808830, "raw_average_value_size": 2181, "num_data_blocks": 140, "num_entries": 829, "num_filter_entries": 829, "num_deletions": 263, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237434, "oldest_key_time": 1764237434, "file_creation_time": 1764237469, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 32, "seqno_to_time_mapping": "N/A"}} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 15] Flush lasted 22256 microseconds, and 7449 cpu microseconds. Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.368256) [db/flush_job.cc:967] [default] [JOB 15] Level-0 flush table #32: 1830127 bytes OK Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.368287) [db/memtable_list.cc:519] [default] Level-0 commit table #32 started Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.370837) [db/memtable_list.cc:722] [default] Level-0 commit table #32: memtable #1 done Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.370861) EVENT_LOG_v1 {"time_micros": 1764237469370854, "job": 15, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.370890) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 15] Try to delete WAL files size 2568161, prev total WAL file size 2568161, number of live WAL files 2. Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000028.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.371791) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6B760031323732' seq:72057594037927935, type:22 .. '6B760031353334' seq:0, type:0; will stop at (end) Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 16] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 15 Base level 0, inputs: [32(1787KB)], [30(18MB)] Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469371895, "job": 16, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [32], "files_L6": [30], "score": -1, "input_data_size": 20975265, "oldest_snapshot_seqno": -1} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 16] Generated table #33: 11204 keys, 20043567 bytes, temperature: kUnknown Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469510492, "cf_name": "default", "job": 16, "event": "table_file_creation", "file_number": 33, "file_size": 20043567, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19978174, "index_size": 36300, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 28037, "raw_key_size": 301213, "raw_average_key_size": 26, "raw_value_size": 19785372, "raw_average_value_size": 1765, "num_data_blocks": 1376, "num_entries": 11204, "num_filter_entries": 11204, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237163, "oldest_key_time": 0, "file_creation_time": 1764237469, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e15a37f6-4d8e-4c18-a493-be6024e004ec", "db_session_id": "GVMC457KA1L6MF3IGLXZ", "orig_file_number": 33, "seqno_to_time_mapping": "N/A"}} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.511012) [db/compaction/compaction_job.cc:1663] [default] [JOB 16] Compacted 1@0 + 1@6 files to L6 => 20043567 bytes Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.522097) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 151.0 rd, 144.3 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(1.7, 18.3 +0.0 blob) out(19.1 +0.0 blob), read-write-amplify(22.4) write-amplify(11.0) OK, records in: 11739, records dropped: 535 output_compression: NoCompression Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.522132) EVENT_LOG_v1 {"time_micros": 1764237469522115, "job": 16, "event": "compaction_finished", "compaction_time_micros": 138901, "compaction_time_cpu_micros": 40982, "output_level": 6, "num_output_files": 1, "total_output_size": 20043567, "num_input_records": 11739, "num_output_records": 11204, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000032.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469522986, "job": 16, "event": "table_file_deletion", "file_number": 32} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000030.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237469525776, "job": 16, "event": "table_file_deletion", "file_number": 30} Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.371689) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.526011) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.526022) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.526025) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.526028) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost ceph-mon[291772]: rocksdb: (Original Log Time 2025/11/27-09:57:49.526032) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:57:49 localhost python3[303813]: ansible-ansible.legacy.command Invoked with _raw_params=ping -W1 -c 3 172.18.0.105 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 04:57:49 localhost nova_compute[284026]: 2025-11-27 09:57:49.973 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:50 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:50 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:50 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:50 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:50 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "osd.4"} v 0) Nov 27 04:57:50 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:57:50 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:50 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:50 localhost ceph-mon[291772]: Reconfiguring osd.1 (monmap changed)... Nov 27 04:57:50 localhost ceph-mon[291772]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:57:50 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:50 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:50 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:57:51 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:51 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:51 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:51 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:51 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} v 0) Nov 27 04:57:51 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:51 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:51 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:51 localhost ceph-mon[291772]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:57:51 localhost ceph-mon[291772]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:57:51 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:51 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:51 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:57:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 7800.1 total, 600.0 interval#012Cumulative writes: 5132 writes, 23K keys, 5132 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5132 writes, 626 syncs, 8.20 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 93 writes, 356 keys, 93 commit groups, 1.0 writes per commit group, ingest: 0.61 MB, 0.00 MB/s#012Interval WAL: 93 writes, 36 syncs, 2.58 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:57:51 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "mgr services"} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mgr services"} : dispatch Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:52 localhost ceph-mon[291772]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:57:52 localhost ceph-mon[291772]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:57:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:57:52 localhost nova_compute[284026]: 2025-11-27 09:57:52.837 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:57:52 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "mon."} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config get", "who": "mon", "key": "public_network"} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch Nov 27 04:57:52 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:52 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:53 localhost podman[303832]: 2025-11-27 09:57:53.00938495 +0000 UTC m=+0.096068438 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 04:57:53 localhost podman[303833]: 2025-11-27 09:57:53.057382808 +0000 UTC m=+0.141514967 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, distribution-scope=public, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, vendor=Red Hat, Inc., version=9.6, release=1755695350, com.redhat.component=ubi9-minimal-container, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.buildah.version=1.33.7, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, vcs-type=git, config_id=edpm, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, build-date=2025-08-20T13:12:41) Nov 27 04:57:53 localhost podman[303833]: 2025-11-27 09:57:53.076960483 +0000 UTC m=+0.161092662 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, managed_by=edpm_ansible, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm, version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, distribution-scope=public, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., architecture=x86_64) Nov 27 04:57:53 localhost podman[303832]: 2025-11-27 09:57:53.077662091 +0000 UTC m=+0.164345589 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible) Nov 27 04:57:53 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/spec.mon}] v 0) Nov 27 04:57:53 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:57:53 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:57:53 localhost ceph-mon[291772]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:57:53 localhost ceph-mon[291772]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:57:53 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:57:53 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:57:53 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:57:53 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:53 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} v 0) Nov 27 04:57:53 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:53 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:53 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:54 localhost ceph-mon[291772]: mon.np0005537446@0(leader).osd e87 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:57:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:57:54 localhost ceph-mon[291772]: Reconfiguring mon.np0005537445 (monmap changed)... Nov 27 04:57:54 localhost ceph-mon[291772]: Reconfiguring daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:57:54 localhost ceph-mon[291772]: Saving service mon spec with placement label:mon Nov 27 04:57:54 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:54 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:54 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:57:54 localhost podman[303919]: 2025-11-27 09:57:54.503683937 +0000 UTC m=+0.099691894 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:57:54 localhost podman[303927]: Nov 27 04:57:54 localhost podman[303919]: 2025-11-27 09:57:54.516131482 +0000 UTC m=+0.112139429 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:57:54 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.571057355 +0000 UTC m=+0.142227346 container create f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, vcs-type=git, vendor=Red Hat, Inc., GIT_CLEAN=True, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, ceph=True, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, RELEASE=main) Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.487750681 +0000 UTC m=+0.058920722 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:54 localhost systemd[1]: Started libpod-conmon-f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398.scope. Nov 27 04:57:54 localhost systemd[1]: Started libcrun container. Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.654584786 +0000 UTC m=+0.225754787 container init f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, version=7, release=553, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., ceph=True, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, RELEASE=main, maintainer=Guillaume Abrioux , architecture=x86_64, vcs-type=git, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.668549 +0000 UTC m=+0.239719001 container start f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, RELEASE=main, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, vcs-type=git, CEPH_POINT_RELEASE=, GIT_CLEAN=True, ceph=True, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.668931791 +0000 UTC m=+0.240101832 container attach f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, architecture=x86_64, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., version=7, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, maintainer=Guillaume Abrioux , release=553, RELEASE=main, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=) Nov 27 04:57:54 localhost vigorous_shamir[303958]: 167 167 Nov 27 04:57:54 localhost systemd[1]: libpod-f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398.scope: Deactivated successfully. Nov 27 04:57:54 localhost podman[303927]: 2025-11-27 09:57:54.674361906 +0000 UTC m=+0.245531897 container died f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, io.openshift.tags=rhceph ceph, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, vendor=Red Hat, Inc., version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, io.openshift.expose-services=, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux ) Nov 27 04:57:54 localhost podman[303963]: 2025-11-27 09:57:54.783227877 +0000 UTC m=+0.096236022 container remove f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=vigorous_shamir, version=7, architecture=x86_64, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, distribution-scope=public, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:57:54 localhost systemd[1]: libpod-conmon-f91685f2890322cef27704761217f4d9f86bd7214e8c80c0afa2c6b43f6b7398.scope: Deactivated successfully. Nov 27 04:57:54 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:57:54 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:54 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:57:54 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:54 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "auth get", "entity": "osd.0"} v 0) Nov 27 04:57:54 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:57:54 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:57:54 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:57:55 localhost nova_compute[284026]: 2025-11-27 09:57:55.001 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:55 localhost ceph-mon[291772]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:57:55 localhost ceph-mon[291772]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:57:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:57:55 localhost ceph-mon[291772]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:57:55 localhost systemd[1]: var-lib-containers-storage-overlay-05fb1fd19372d64c6e8dc98cb3dbcd27fdb7f4eceb8fb8def7851acb9804e403-merged.mount: Deactivated successfully. Nov 27 04:57:55 localhost podman[304030]: Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.580737872 +0000 UTC m=+0.093341235 container create 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, name=rhceph, GIT_BRANCH=main, architecture=x86_64, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, RELEASE=main, ceph=True, release=553, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , distribution-scope=public, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:55 localhost systemd[1]: Started libpod-conmon-148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b.scope. Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.544565492 +0000 UTC m=+0.057168875 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:55 localhost systemd[1]: Started libcrun container. Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.659573047 +0000 UTC m=+0.172176410 container init 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, RELEASE=main, version=7, CEPH_POINT_RELEASE=, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, release=553, GIT_BRANCH=main, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, vcs-type=git) Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.681946277 +0000 UTC m=+0.194549610 container start 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, architecture=x86_64, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, GIT_BRANCH=main, io.buildah.version=1.33.12, RELEASE=main, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, release=553, version=7, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.682165893 +0000 UTC m=+0.194769226 container attach 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, release=553, com.redhat.component=rhceph-container, architecture=x86_64, version=7, io.openshift.expose-services=, vcs-type=git, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, GIT_BRANCH=main, vendor=Red Hat, Inc.) Nov 27 04:57:55 localhost zen_lamarr[304045]: 167 167 Nov 27 04:57:55 localhost systemd[1]: libpod-148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b.scope: Deactivated successfully. Nov 27 04:57:55 localhost podman[304030]: 2025-11-27 09:57:55.68465519 +0000 UTC m=+0.197258533 container died 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, version=7, architecture=x86_64, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, release=553, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, RELEASE=main, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, name=rhceph, vcs-type=git, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, vendor=Red Hat, Inc., GIT_CLEAN=True, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:57:55 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "quorum_status"} v 0) Nov 27 04:57:55 localhost ceph-mon[291772]: log_channel(audit) log [DBG] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "quorum_status"} : dispatch Nov 27 04:57:55 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e15 handle_command mon_command({"prefix": "mon rm", "name": "np0005537446"} v 0) Nov 27 04:57:55 localhost ceph-mon[291772]: log_channel(audit) log [INF] : from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "mon rm", "name": "np0005537446"} : dispatch Nov 27 04:57:55 localhost openstack_network_exporter[244641]: ERROR 09:57:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:57:55 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x561963d90160 mon_map magic: 0 from mon.0 v2:172.18.0.108:3300/0 Nov 27 04:57:55 localhost openstack_network_exporter[244641]: ERROR 09:57:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:57:55 localhost openstack_network_exporter[244641]: ERROR 09:57:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:57:55 localhost openstack_network_exporter[244641]: ERROR 09:57:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:57:55 localhost openstack_network_exporter[244641]: Nov 27 04:57:55 localhost openstack_network_exporter[244641]: ERROR 09:57:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:57:55 localhost openstack_network_exporter[244641]: Nov 27 04:57:55 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.103:3300/0 Nov 27 04:57:55 localhost ceph-mgr[290377]: client.0 ms_handle_reset on v2:172.18.0.103:3300/0 Nov 27 04:57:55 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x56195ce59600 mon_map magic: 0 from mon.0 v2:172.18.0.103:3300/0 Nov 27 04:57:55 localhost ceph-mon[291772]: mon.np0005537446@0(leader) e16 removed from monmap, suicide. Nov 27 04:57:55 localhost podman[304050]: 2025-11-27 09:57:55.796375877 +0000 UTC m=+0.102232304 container remove 148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=zen_lamarr, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, CEPH_POINT_RELEASE=, GIT_CLEAN=True, io.buildah.version=1.33.12, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux ) Nov 27 04:57:55 localhost systemd[1]: libpod-conmon-148d250105a39f46c61e98734908d17414c276575a7f6d299480cbd58b2e486b.scope: Deactivated successfully. Nov 27 04:57:55 localhost podman[304081]: 2025-11-27 09:57:55.903980573 +0000 UTC m=+0.062231390 container died f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, GIT_BRANCH=main, io.openshift.expose-services=, distribution-scope=public, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, io.buildah.version=1.33.12, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, release=553, name=rhceph) Nov 27 04:57:55 localhost podman[304081]: 2025-11-27 09:57:55.944550062 +0000 UTC m=+0.102800849 container remove f1873e03658e60918b0f16e299f9483f4cfb36055a6f0933562ac75c119deb4c (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , version=7, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, io.openshift.expose-services=, distribution-scope=public, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., ceph=True, architecture=x86_64, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, vcs-type=git, name=rhceph, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:57:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 7800.2 total, 600.0 interval#012Cumulative writes: 5791 writes, 25K keys, 5791 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.00 MB/s#012Cumulative WAL: 5791 writes, 847 syncs, 6.84 writes per sync, written: 0.02 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 198 writes, 450 keys, 198 commit groups, 1.0 writes per commit group, ingest: 0.40 MB, 0.00 MB/s#012Interval WAL: 198 writes, 94 syncs, 2.11 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 04:57:56 localhost systemd[1]: tmp-crun.skjNI8.mount: Deactivated successfully. Nov 27 04:57:56 localhost systemd[1]: var-lib-containers-storage-overlay-a7e282e1ae4ba490a6a5444e259ffed76d99923d61fc4dfd3a156e11a6d53110-merged.mount: Deactivated successfully. Nov 27 04:57:56 localhost systemd[1]: var-lib-containers-storage-overlay-d2767724e77723f2921128140e380c6e7b2790b261ea5617ea23aa84e7c9c784-merged.mount: Deactivated successfully. Nov 27 04:57:57 localhost systemd[1]: ceph-e83f3b0c-4090-52df-95d4-ad9be8516692@mon.np0005537446.service: Deactivated successfully. Nov 27 04:57:57 localhost systemd[1]: Stopped Ceph mon.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 04:57:57 localhost systemd[1]: ceph-e83f3b0c-4090-52df-95d4-ad9be8516692@mon.np0005537446.service: Consumed 12.781s CPU time. Nov 27 04:57:57 localhost systemd[1]: Reloading. Nov 27 04:57:57 localhost systemd-sysv-generator[304305]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:57:57 localhost systemd-rc-local-generator[304302]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:57:57 localhost podman[304315]: Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.696310146 +0000 UTC m=+0.085191475 container create 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, GIT_CLEAN=True, distribution-scope=public, version=7, vendor=Red Hat, Inc., maintainer=Guillaume Abrioux , io.openshift.expose-services=, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, ceph=True, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:57:57 localhost systemd[1]: Started libpod-conmon-723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa.scope. Nov 27 04:57:57 localhost systemd[1]: Started libcrun container. Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.661101081 +0000 UTC m=+0.049982430 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.766021567 +0000 UTC m=+0.154902856 container init 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, version=7, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, vcs-type=git, name=rhceph, io.buildah.version=1.33.12, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container) Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.775069519 +0000 UTC m=+0.163950808 container start 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, version=7, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, architecture=x86_64, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, com.redhat.component=rhceph-container, RELEASE=main, name=rhceph, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12) Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.775195103 +0000 UTC m=+0.164076392 container attach 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, ceph=True, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, version=7) Nov 27 04:57:57 localhost stupefied_hertz[304330]: 167 167 Nov 27 04:57:57 localhost systemd[1]: libpod-723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa.scope: Deactivated successfully. Nov 27 04:57:57 localhost podman[304315]: 2025-11-27 09:57:57.782732775 +0000 UTC m=+0.171614124 container died 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, maintainer=Guillaume Abrioux , GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, release=553, vcs-type=git, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, io.openshift.tags=rhceph ceph, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, name=rhceph, version=7, vendor=Red Hat, Inc., GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:57:57 localhost nova_compute[284026]: 2025-11-27 09:57:57.882 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:57:57 localhost podman[304335]: 2025-11-27 09:57:57.920553742 +0000 UTC m=+0.126871214 container remove 723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=stupefied_hertz, description=Red Hat Ceph Storage 7, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, ceph=True, maintainer=Guillaume Abrioux , GIT_BRANCH=main, distribution-scope=public, io.openshift.expose-services=, vcs-type=git, release=553, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, name=rhceph, GIT_CLEAN=True, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., version=7, CEPH_POINT_RELEASE=) Nov 27 04:57:57 localhost systemd[1]: libpod-conmon-723e143e6d964865d2a7a193e52feb908cc66846ad3a93cf0d1eafc84816e0fa.scope: Deactivated successfully. Nov 27 04:57:58 localhost systemd[1]: var-lib-containers-storage-overlay-e695b978487798803ec78541845ec5433898313df7707ecb3b41e3d92eaef793-merged.mount: Deactivated successfully. Nov 27 04:57:58 localhost podman[304412]: Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.846512734 +0000 UTC m=+0.082339090 container create 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, name=rhceph, release=553, vcs-type=git, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:57:58 localhost systemd[1]: Started libpod-conmon-8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f.scope. Nov 27 04:57:58 localhost systemd[1]: Started libcrun container. Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.81170338 +0000 UTC m=+0.047529826 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.919706977 +0000 UTC m=+0.155533343 container init 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, io.k8s.description=Red Hat Ceph Storage 7, release=553, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, vcs-type=git, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, vendor=Red Hat, Inc., io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, RELEASE=main, GIT_CLEAN=True, distribution-scope=public, ceph=True) Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.930223269 +0000 UTC m=+0.166049635 container start 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, description=Red Hat Ceph Storage 7, ceph=True, RELEASE=main, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., version=7, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, name=rhceph, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, release=553, io.openshift.expose-services=, distribution-scope=public, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.930580819 +0000 UTC m=+0.166407185 container attach 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, release=553, version=7, com.redhat.component=rhceph-container, name=rhceph, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., ceph=True, vcs-type=git, maintainer=Guillaume Abrioux , io.openshift.expose-services=, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:57:58 localhost musing_ardinghelli[304428]: 167 167 Nov 27 04:57:58 localhost systemd[1]: libpod-8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f.scope: Deactivated successfully. Nov 27 04:57:58 localhost podman[304412]: 2025-11-27 09:57:58.934184266 +0000 UTC m=+0.170010652 container died 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, RELEASE=main, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, maintainer=Guillaume Abrioux , release=553, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, vcs-type=git, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main) Nov 27 04:57:59 localhost podman[304433]: 2025-11-27 09:57:59.040678183 +0000 UTC m=+0.093850740 container remove 8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=musing_ardinghelli, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., GIT_CLEAN=True, RELEASE=main, com.redhat.component=rhceph-container, vcs-type=git, release=553, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, architecture=x86_64, distribution-scope=public, io.openshift.tags=rhceph ceph) Nov 27 04:57:59 localhost systemd[1]: libpod-conmon-8c645590515b65caa412fbb29eba07b8c5b211eae5bf16f51f81fe83c97b2b1f.scope: Deactivated successfully. Nov 27 04:57:59 localhost systemd[1]: var-lib-containers-storage-overlay-aa5fd2c6beda4d40f9763a240f266812a7b168d9709701896ef7c6e79d4552fb-merged.mount: Deactivated successfully. Nov 27 04:57:59 localhost podman[304503]: Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.828039655 +0000 UTC m=+0.085900606 container create 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, vendor=Red Hat, Inc., GIT_BRANCH=main, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , ceph=True, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, architecture=x86_64, io.buildah.version=1.33.12, name=rhceph, vcs-type=git, RELEASE=main, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:57:59 localhost systemd[1]: Started libpod-conmon-50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9.scope. Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.795843611 +0000 UTC m=+0.053704612 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:57:59 localhost systemd[1]: Started libcrun container. Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.911841973 +0000 UTC m=+0.169702944 container init 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., name=rhceph, maintainer=Guillaume Abrioux , RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, version=7, description=Red Hat Ceph Storage 7, architecture=x86_64, io.buildah.version=1.33.12, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55) Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.92700974 +0000 UTC m=+0.184870711 container start 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, maintainer=Guillaume Abrioux , vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, ceph=True, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, version=7, architecture=x86_64, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, distribution-scope=public, CEPH_POINT_RELEASE=, release=553, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, RELEASE=main) Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.927274707 +0000 UTC m=+0.185135678 container attach 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, name=rhceph, io.openshift.tags=rhceph ceph, RELEASE=main, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, ceph=True, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d) Nov 27 04:57:59 localhost determined_lumiere[304519]: 167 167 Nov 27 04:57:59 localhost systemd[1]: libpod-50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9.scope: Deactivated successfully. Nov 27 04:57:59 localhost podman[304503]: 2025-11-27 09:57:59.932280891 +0000 UTC m=+0.190141902 container died 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, RELEASE=main, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, architecture=x86_64, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, name=rhceph, release=553, io.openshift.expose-services=) Nov 27 04:58:00 localhost nova_compute[284026]: 2025-11-27 09:58:00.004 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:00 localhost podman[304524]: 2025-11-27 09:58:00.040570387 +0000 UTC m=+0.098169455 container remove 50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=determined_lumiere, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, GIT_BRANCH=main, vendor=Red Hat, Inc., io.openshift.expose-services=, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, version=7, distribution-scope=public, RELEASE=main, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, release=553) Nov 27 04:58:00 localhost systemd[1]: libpod-conmon-50b61b538a5302906a09d55a3ee74e8e83e1f90b0e6ea7b261fa88e7efe8f0c9.scope: Deactivated successfully. Nov 27 04:58:00 localhost systemd[1]: tmp-crun.Y1UJMd.mount: Deactivated successfully. Nov 27 04:58:00 localhost systemd[1]: var-lib-containers-storage-overlay-e9e45a8c87fd88bae93fcfeb9a4bf93d09af332223308dab319886732139ff42-merged.mount: Deactivated successfully. Nov 27 04:58:01 localhost podman[304647]: 2025-11-27 09:58:01.266079233 +0000 UTC m=+0.092860242 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, version=7, io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, release=553, io.openshift.tags=rhceph ceph, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., io.openshift.expose-services=, GIT_BRANCH=main, name=rhceph, architecture=x86_64, maintainer=Guillaume Abrioux , ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 04:58:01 localhost podman[304647]: 2025-11-27 09:58:01.383242046 +0000 UTC m=+0.210023045 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, release=553, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, ceph=True, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, vcs-type=git, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc.) Nov 27 04:58:02 localhost nova_compute[284026]: 2025-11-27 09:58:02.929 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:58:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:58:04 localhost podman[305064]: 2025-11-27 09:58:04.261645646 +0000 UTC m=+0.116012644 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:58:04 localhost podman[305064]: 2025-11-27 09:58:04.272012503 +0000 UTC m=+0.126379472 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:58:04 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:58:04 localhost podman[305114]: 2025-11-27 09:58:04.370172347 +0000 UTC m=+0.097106936 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible) Nov 27 04:58:04 localhost podman[305114]: 2025-11-27 09:58:04.484430632 +0000 UTC m=+0.211365201 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 04:58:04 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:58:05 localhost nova_compute[284026]: 2025-11-27 09:58:05.007 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:07 localhost nova_compute[284026]: 2025-11-27 09:58:07.964 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:08 localhost podman[242678]: time="2025-11-27T09:58:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:58:08 localhost podman[242678]: @ - - [27/Nov/2025:09:58:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 155317 "" "Go-http-client/1.1" Nov 27 04:58:08 localhost podman[242678]: @ - - [27/Nov/2025:09:58:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19218 "" "Go-http-client/1.1" Nov 27 04:58:09 localhost podman[305279]: Nov 27 04:58:09 localhost podman[305279]: 2025-11-27 09:58:09.943824212 +0000 UTC m=+0.092770350 container create 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , RELEASE=main, distribution-scope=public, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_CLEAN=True, name=rhceph, release=553, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12) Nov 27 04:58:09 localhost systemd[1]: Started libpod-conmon-4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35.scope. Nov 27 04:58:10 localhost podman[305279]: 2025-11-27 09:58:09.899979875 +0000 UTC m=+0.048926053 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:10 localhost systemd[1]: Started libcrun container. Nov 27 04:58:10 localhost nova_compute[284026]: 2025-11-27 09:58:10.009 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:10 localhost podman[305279]: 2025-11-27 09:58:10.025641937 +0000 UTC m=+0.174588075 container init 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, distribution-scope=public, GIT_BRANCH=main, release=553, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, description=Red Hat Ceph Storage 7, ceph=True, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:58:10 localhost systemd[1]: tmp-crun.HvEDXL.mount: Deactivated successfully. Nov 27 04:58:10 localhost podman[305279]: 2025-11-27 09:58:10.04516473 +0000 UTC m=+0.194110868 container start 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, RELEASE=main, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, GIT_BRANCH=main, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, com.redhat.component=rhceph-container, vcs-type=git, io.openshift.expose-services=, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, GIT_CLEAN=True, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:58:10 localhost podman[305279]: 2025-11-27 09:58:10.048349186 +0000 UTC m=+0.197295374 container attach 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, ceph=True, io.openshift.tags=rhceph ceph, name=rhceph, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, maintainer=Guillaume Abrioux , distribution-scope=public, version=7, io.buildah.version=1.33.12, architecture=x86_64, GIT_CLEAN=True, RELEASE=main) Nov 27 04:58:10 localhost exciting_mcnulty[305294]: 167 167 Nov 27 04:58:10 localhost systemd[1]: libpod-4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35.scope: Deactivated successfully. Nov 27 04:58:10 localhost podman[305279]: 2025-11-27 09:58:10.053499244 +0000 UTC m=+0.202445402 container died 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, ceph=True, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, name=rhceph, release=553, RELEASE=main, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, distribution-scope=public, vcs-type=git, GIT_BRANCH=main, version=7, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph) Nov 27 04:58:10 localhost podman[305299]: 2025-11-27 09:58:10.157454853 +0000 UTC m=+0.094784264 container remove 4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=exciting_mcnulty, distribution-scope=public, architecture=x86_64, GIT_BRANCH=main, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, name=rhceph, description=Red Hat Ceph Storage 7, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, vcs-type=git, ceph=True, io.openshift.expose-services=) Nov 27 04:58:10 localhost systemd[1]: libpod-conmon-4fa3ce694fcbd56f26fa197e27b2fcaba7f75c9f9fe9a44c48e36e33b20d0a35.scope: Deactivated successfully. Nov 27 04:58:10 localhost podman[305316]: Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.284357028 +0000 UTC m=+0.086129702 container create 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., name=rhceph, vcs-type=git, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, version=7, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, ceph=True, release=553, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, io.openshift.expose-services=) Nov 27 04:58:10 localhost systemd[1]: Started libpod-conmon-74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095.scope. Nov 27 04:58:10 localhost systemd[1]: Started libcrun container. Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.248870436 +0000 UTC m=+0.050643150 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3dae10b7e6e6a5c98b699dba651589d568c722ebd4018e92b4eb87f7646a2c83/merged/tmp/config supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3dae10b7e6e6a5c98b699dba651589d568c722ebd4018e92b4eb87f7646a2c83/merged/tmp/keyring supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3dae10b7e6e6a5c98b699dba651589d568c722ebd4018e92b4eb87f7646a2c83/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:10 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/3dae10b7e6e6a5c98b699dba651589d568c722ebd4018e92b4eb87f7646a2c83/merged/var/lib/ceph/mon/ceph-np0005537446 supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.357906161 +0000 UTC m=+0.159678835 container init 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, version=7, io.k8s.description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, ceph=True, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, name=rhceph, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., vcs-type=git, build-date=2025-09-24T08:57:55, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, architecture=x86_64, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, description=Red Hat Ceph Storage 7) Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.368621428 +0000 UTC m=+0.170394102 container start 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, GIT_CLEAN=True, io.openshift.expose-services=, name=rhceph, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, vcs-type=git, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, architecture=x86_64, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, RELEASE=main, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.368941227 +0000 UTC m=+0.170713941 container attach 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, version=7, RELEASE=main, ceph=True, vcs-type=git, CEPH_POINT_RELEASE=) Nov 27 04:58:10 localhost systemd[1]: libpod-74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095.scope: Deactivated successfully. Nov 27 04:58:10 localhost podman[305316]: 2025-11-27 09:58:10.480382937 +0000 UTC m=+0.282155621 container died 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, architecture=x86_64, com.redhat.component=rhceph-container, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, RELEASE=main, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph) Nov 27 04:58:10 localhost podman[305357]: 2025-11-27 09:58:10.580710317 +0000 UTC m=+0.087930229 container remove 74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_hertz, io.buildah.version=1.33.12, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, distribution-scope=public, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, architecture=x86_64, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, ceph=True, vcs-type=git, com.redhat.component=rhceph-container, build-date=2025-09-24T08:57:55, io.k8s.description=Red Hat Ceph Storage 7, version=7, RELEASE=main) Nov 27 04:58:10 localhost systemd[1]: libpod-conmon-74620ae601ab357fd3a8e6d428e12aa500f5648c82b08e27601871343830d095.scope: Deactivated successfully. Nov 27 04:58:10 localhost systemd[1]: Reloading. Nov 27 04:58:10 localhost systemd-sysv-generator[305401]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:58:10 localhost systemd-rc-local-generator[305397]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:10 localhost systemd[1]: var-lib-containers-storage-overlay-f180de65205c6240a3a64398b30c2458c1d5fbdce852f5cba4e7564822cc3938-merged.mount: Deactivated successfully. Nov 27 04:58:11 localhost systemd[1]: Reloading. Nov 27 04:58:11 localhost systemd-rc-local-generator[305437]: /etc/rc.d/rc.local is not marked executable, skipping. Nov 27 04:58:11 localhost systemd-sysv-generator[305443]: SysV service '/etc/rc.d/init.d/network' lacks a native systemd unit file. Automatically generating a unit file for compatibility. Please update package to include a native systemd unit file, in order to make it more safe and robust. Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtsecretd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtqemud.service:25: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtproxyd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtnodedevd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/insights-client-boot.service:24: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtstoraged.service:20: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtnwfilterd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtnetworkd.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: /usr/lib/systemd/system/virtinterfaced.service:18: Failed to parse service type, ignoring: notify-reload Nov 27 04:58:11 localhost systemd[1]: Starting Ceph mon.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692... Nov 27 04:58:11 localhost podman[305502]: Nov 27 04:58:11 localhost podman[305502]: 2025-11-27 09:58:11.776759434 +0000 UTC m=+0.053599849 container create ef73efb04e5804ad281ad2ecd50b982ff03322633f02dd2ab3ed910d2960ca86 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, RELEASE=main, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, com.redhat.component=rhceph-container, architecture=x86_64, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, description=Red Hat Ceph Storage 7, distribution-scope=public, io.buildah.version=1.33.12, vcs-type=git, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:58:11 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/15d88a99950ca8d37edf354058ee41abdc42863521e2a2fd898bbc713b7fc9f7/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:11 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/15d88a99950ca8d37edf354058ee41abdc42863521e2a2fd898bbc713b7fc9f7/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:11 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/15d88a99950ca8d37edf354058ee41abdc42863521e2a2fd898bbc713b7fc9f7/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:11 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/15d88a99950ca8d37edf354058ee41abdc42863521e2a2fd898bbc713b7fc9f7/merged/var/lib/ceph/mon/ceph-np0005537446 supports timestamps until 2038 (0x7fffffff) Nov 27 04:58:11 localhost podman[305502]: 2025-11-27 09:58:11.835204522 +0000 UTC m=+0.112044947 container init ef73efb04e5804ad281ad2ecd50b982ff03322633f02dd2ab3ed910d2960ca86 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, io.openshift.expose-services=, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, GIT_BRANCH=main, version=7, maintainer=Guillaume Abrioux , GIT_CLEAN=True, io.buildah.version=1.33.12, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, architecture=x86_64, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553) Nov 27 04:58:11 localhost podman[305502]: 2025-11-27 09:58:11.843018852 +0000 UTC m=+0.119859277 container start ef73efb04e5804ad281ad2ecd50b982ff03322633f02dd2ab3ed910d2960ca86 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mon-np0005537446, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, version=7, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, io.buildah.version=1.33.12, architecture=x86_64, ceph=True, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, GIT_BRANCH=main, name=rhceph, maintainer=Guillaume Abrioux ) Nov 27 04:58:11 localhost bash[305502]: ef73efb04e5804ad281ad2ecd50b982ff03322633f02dd2ab3ed910d2960ca86 Nov 27 04:58:11 localhost podman[305502]: 2025-11-27 09:58:11.75425754 +0000 UTC m=+0.031097965 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:11 localhost systemd[1]: Started Ceph mon.np0005537446 for e83f3b0c-4090-52df-95d4-ad9be8516692. Nov 27 04:58:11 localhost ceph-mon[305520]: set uid:gid to 167:167 (ceph:ceph) Nov 27 04:58:11 localhost ceph-mon[305520]: ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable), process ceph-mon, pid 2 Nov 27 04:58:11 localhost ceph-mon[305520]: pidfile_write: ignore empty --pid-file Nov 27 04:58:11 localhost ceph-mon[305520]: load: jerasure load: lrc Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: RocksDB version: 7.9.2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Git sha 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Compile date 2025-09-23 00:00:00 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: DB SUMMARY Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: DB Session ID: 538JK4PSKJ851L30ETJB Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: CURRENT file: CURRENT Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: IDENTITY file: IDENTITY Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: SST files in /var/lib/ceph/mon/ceph-np0005537446/store.db dir, Total Num: 0, files: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-np0005537446/store.db: 000004.log size: 636 ; Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.error_if_exists: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.create_if_missing: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.paranoid_checks: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.flush_verify_memtable_count: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.env: 0x55fa184279e0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.fs: PosixFileSystem Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.info_log: 0x55fa1a6d8d20 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_file_opening_threads: 16 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.statistics: (nil) Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.use_fsync: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_log_file_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_manifest_file_size: 1073741824 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.log_file_time_to_roll: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.keep_log_file_num: 1000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.recycle_log_file_num: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_fallocate: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_mmap_reads: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_mmap_writes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.use_direct_reads: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.create_missing_column_families: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.db_log_dir: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.wal_dir: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.table_cache_numshardbits: 6 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.WAL_ttl_seconds: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.WAL_size_limit_MB: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.manifest_preallocation_size: 4194304 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.is_fd_close_on_exec: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.advise_random_on_open: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.db_write_buffer_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.write_buffer_manager: 0x55fa1a6e9540 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.access_hint_on_compaction_start: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.random_access_max_buffer_size: 1048576 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.use_adaptive_mutex: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.rate_limiter: (nil) Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.wal_recovery_mode: 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enable_thread_tracking: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enable_pipelined_write: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.unordered_write: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_concurrent_memtable_write: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.write_thread_max_yield_usec: 100 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.write_thread_slow_yield_usec: 3 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.row_cache: None Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.wal_filter: None Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.avoid_flush_during_recovery: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_ingest_behind: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.two_write_queues: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.manual_wal_flush: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.wal_compression: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.atomic_flush: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.persist_stats_to_disk: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.write_dbid_to_manifest: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.log_readahead_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.file_checksum_gen_factory: Unknown Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.best_efforts_recovery: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bgerror_resume_count: 2147483647 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.allow_data_in_errors: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.db_host_id: __hostname__ Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enforce_single_del_contracts: true Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_background_jobs: 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_background_compactions: -1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_subcompactions: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.avoid_flush_during_shutdown: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.writable_file_max_buffer_size: 1048576 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.delayed_write_rate : 16777216 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_total_wal_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.stats_dump_period_sec: 600 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.stats_persist_period_sec: 600 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.stats_history_buffer_size: 1048576 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_open_files: -1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bytes_per_sync: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.wal_bytes_per_sync: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.strict_bytes_per_sync: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_readahead_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_background_flushes: -1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Compression algorithms supported: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kZSTD supported: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kXpressCompression supported: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kBZip2Compression supported: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kZSTDNotFinalCompression supported: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kLZ4Compression supported: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kZlibCompression supported: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kLZ4HCCompression supported: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: #011kSnappyCompression supported: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Fast CRC32 supported: Supported on x86 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: DMutex implementation: pthread_mutex_t Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-np0005537446/store.db/MANIFEST-000005 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.comparator: leveldb.BytewiseComparator Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.merge_operator: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_filter: None Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_filter_factory: None Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.sst_partitioner_factory: None Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.memtable_factory: SkipListFactory Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.table_factory: BlockBasedTable Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55fa1a6d8980)#012 cache_index_and_filter_blocks: 1#012 cache_index_and_filter_blocks_with_high_priority: 0#012 pin_l0_filter_and_index_blocks_in_cache: 0#012 pin_top_level_index_and_filter: 1#012 index_type: 0#012 data_block_index_type: 0#012 index_shortening: 1#012 data_block_hash_table_util_ratio: 0.750000#012 checksum: 4#012 no_block_cache: 0#012 block_cache: 0x55fa1a6d5350#012 block_cache_name: BinnedLRUCache#012 block_cache_options:#012 capacity : 536870912#012 num_shard_bits : 4#012 strict_capacity_limit : 0#012 high_pri_pool_ratio: 0.000#012 block_cache_compressed: (nil)#012 persistent_cache: (nil)#012 block_size: 4096#012 block_size_deviation: 10#012 block_restart_interval: 16#012 index_block_restart_interval: 1#012 metadata_block_size: 4096#012 partition_filters: 0#012 use_delta_encoding: 1#012 filter_policy: bloomfilter#012 whole_key_filtering: 1#012 verify_compression: 0#012 read_amp_bytes_per_bit: 0#012 format_version: 5#012 enable_index_compression: 1#012 block_align: 0#012 max_auto_readahead_size: 262144#012 prepopulate_block_cache: 0#012 initial_auto_readahead_size: 8192#012 num_file_reads_for_auto_readahead: 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.write_buffer_size: 33554432 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_write_buffer_number: 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression: NoCompression Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression: Disabled Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.prefix_extractor: nullptr Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.num_levels: 7 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.min_write_buffer_number_to_merge: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.level: 32767 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.strategy: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.enabled: false Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.window_bits: -14 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.level: 32767 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.strategy: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.max_dict_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.parallel_threads: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.enabled: false Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.level0_file_num_compaction_trigger: 4 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.level0_slowdown_writes_trigger: 20 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.level0_stop_writes_trigger: 36 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.target_file_size_base: 67108864 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.target_file_size_multiplier: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_base: 268435456 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_sequential_skip_in_iterations: 8 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_compaction_bytes: 1677721600 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.arena_block_size: 1048576 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.disable_auto_compactions: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_style: kCompactionStyleLevel Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_pri: kMinOverlappingRatio Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.size_ratio: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.table_properties_collectors: Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.inplace_update_support: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.inplace_update_num_locks: 10000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.memtable_whole_key_filtering: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.memtable_huge_page_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.bloom_locality: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.max_successive_merges: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.optimize_filters_for_hits: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.paranoid_file_checks: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.force_consistency_checks: 1 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.report_bg_io_stats: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.ttl: 2592000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.periodic_compaction_seconds: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.preclude_last_level_data_seconds: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.preserve_internal_time_seconds: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enable_blob_files: false Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.min_blob_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_file_size: 268435456 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_compression_type: NoCompression Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.enable_blob_garbage_collection: false Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_compaction_readahead_size: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.blob_file_starting_level: 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-np0005537446/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: a7e2bcde-7690-4ed3-a926-bff021ee997e Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237491894913, "job": 1, "event": "recovery_started", "wal_files": [4]} Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237491897510, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1762, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 648, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 526, "raw_average_value_size": 105, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237491897716, "job": 1, "event": "recovery_finished"} Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55fa1a6fce00 Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: DB pointer 0x55fa1a7f2000 Nov 27 04:58:11 localhost ceph-mon[305520]: mon.np0005537446 does not exist in monmap, will attempt to join an existing cluster Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 04:58:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 0.0 total, 0.0 interval#012Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s#012Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s#012Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 1/0 1.72 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012 Sum 1/0 1.72 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 0.0 total, 0.0 interval#012Flush(GB): cumulative 0.000, interval 0.000#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.00 GB write, 0.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Interval compaction: 0.00 GB write, 0.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x55fa1a6d5350#2 capacity: 512.00 MB usage: 0.22 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.6e-05 secs_since: 0#012Block cache entry stats(count,size,portion): FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] ** Nov 27 04:58:11 localhost ceph-mon[305520]: using public_addr v2:172.18.0.105:0/0 -> [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] Nov 27 04:58:11 localhost ceph-mon[305520]: starting mon.np0005537446 rank -1 at public addrs [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] at bind addrs [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] mon_data /var/lib/ceph/mon/ceph-np0005537446 fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:58:11 localhost ceph-mon[305520]: mon.np0005537446@-1(???) e0 preinit fsid e83f3b0c-4090-52df-95d4-ad9be8516692 Nov 27 04:58:12 localhost nova_compute[284026]: 2025-11-27 09:58:12.753 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:12 localhost nova_compute[284026]: 2025-11-27 09:58:12.754 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:12 localhost nova_compute[284026]: 2025-11-27 09:58:12.754 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:13 localhost nova_compute[284026]: 2025-11-27 09:58:13.023 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:13 localhost nova_compute[284026]: 2025-11-27 09:58:13.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:13 localhost nova_compute[284026]: 2025-11-27 09:58:13.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:58:14 localhost podman[305612]: Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.42393816 +0000 UTC m=+0.082024951 container create b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, vcs-type=git, io.openshift.tags=rhceph ceph, io.openshift.expose-services=, maintainer=Guillaume Abrioux , version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, vendor=Red Hat, Inc., RELEASE=main, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, io.buildah.version=1.33.12) Nov 27 04:58:14 localhost systemd[1]: Started libpod-conmon-b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05.scope. Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.388256692 +0000 UTC m=+0.046343483 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:14 localhost systemd[1]: Started libcrun container. Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.512051444 +0000 UTC m=+0.170138235 container init b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, RELEASE=main, ceph=True, vcs-type=git, name=rhceph, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, architecture=x86_64, release=553) Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.524017235 +0000 UTC m=+0.182104036 container start b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, version=7, vendor=Red Hat, Inc., io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, maintainer=Guillaume Abrioux , RELEASE=main, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, ceph=True, architecture=x86_64, io.openshift.tags=rhceph ceph, release=553, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7) Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.524343784 +0000 UTC m=+0.182430585 container attach b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, architecture=x86_64, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , RELEASE=main, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., io.openshift.expose-services=, distribution-scope=public, vcs-type=git, name=rhceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, GIT_CLEAN=True) Nov 27 04:58:14 localhost thirsty_ride[305627]: 167 167 Nov 27 04:58:14 localhost systemd[1]: libpod-b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05.scope: Deactivated successfully. Nov 27 04:58:14 localhost podman[305612]: 2025-11-27 09:58:14.528277279 +0000 UTC m=+0.186364070 container died b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , version=7, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, release=553, architecture=x86_64, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.buildah.version=1.33.12, RELEASE=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:58:14 localhost podman[305634]: 2025-11-27 09:58:14.624694666 +0000 UTC m=+0.085856065 container remove b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=thirsty_ride, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, ceph=True, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, architecture=x86_64, vcs-type=git, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., maintainer=Guillaume Abrioux ) Nov 27 04:58:14 localhost systemd[1]: libpod-conmon-b3931ba27d26b3d469cd1b87083ecbbd13971c4d862d9d835401e64d722c3a05.scope: Deactivated successfully. Nov 27 04:58:14 localhost nova_compute[284026]: 2025-11-27 09:58:14.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:15 localhost nova_compute[284026]: 2025-11-27 09:58:15.015 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:15 localhost podman[305700]: Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.380102602 +0000 UTC m=+0.078482947 container create 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, GIT_BRANCH=main, io.openshift.expose-services=, GIT_CLEAN=True, version=7, vcs-type=git, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, release=553, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, ceph=True, io.buildah.version=1.33.12, vendor=Red Hat, Inc., RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:58:15 localhost systemd[1]: Started libpod-conmon-7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92.scope. Nov 27 04:58:15 localhost systemd[1]: var-lib-containers-storage-overlay-5dc9744cba9f64d7d26c87a21a9c182e438826c9b9325a9377d811dced14fef3-merged.mount: Deactivated successfully. Nov 27 04:58:15 localhost systemd[1]: Started libcrun container. Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.350386104 +0000 UTC m=+0.048766479 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.452708319 +0000 UTC m=+0.151088674 container init 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vcs-type=git, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, io.openshift.expose-services=, GIT_BRANCH=main, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, io.buildah.version=1.33.12, version=7, build-date=2025-09-24T08:57:55, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.463054266 +0000 UTC m=+0.161434611 container start 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., release=553, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.buildah.version=1.33.12, GIT_REPO=https://github.com/ceph/ceph-container.git, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, io.openshift.tags=rhceph ceph, vcs-type=git, architecture=x86_64, name=rhceph, GIT_BRANCH=main, RELEASE=main, distribution-scope=public, version=7, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux ) Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.46355351 +0000 UTC m=+0.161933855 container attach 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, vendor=Red Hat, Inc., io.openshift.expose-services=, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, distribution-scope=public, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, RELEASE=main, version=7, com.redhat.component=rhceph-container, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55) Nov 27 04:58:15 localhost systemd[1]: libpod-7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92.scope: Deactivated successfully. Nov 27 04:58:15 localhost serene_jackson[305716]: 167 167 Nov 27 04:58:15 localhost podman[305700]: 2025-11-27 09:58:15.466921471 +0000 UTC m=+0.165301866 container died 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, build-date=2025-09-24T08:57:55, release=553, ceph=True, io.buildah.version=1.33.12, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, distribution-scope=public, description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , version=7, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, architecture=x86_64, io.openshift.expose-services=, RELEASE=main, io.openshift.tags=rhceph ceph) Nov 27 04:58:15 localhost podman[305721]: 2025-11-27 09:58:15.56980371 +0000 UTC m=+0.095011700 container remove 7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=serene_jackson, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.tags=rhceph ceph, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, name=rhceph, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, release=553, vendor=Red Hat, Inc., ceph=True, io.openshift.expose-services=, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, GIT_BRANCH=main) Nov 27 04:58:15 localhost systemd[1]: libpod-conmon-7aeb1e5eb936a2ae6b3100356df1749e2b9571389e159fabee35ee73b3991e92.scope: Deactivated successfully. Nov 27 04:58:16 localhost systemd[1]: var-lib-containers-storage-overlay-1853546476e44e7315100405a72c7a547e3f6a756efcb582637e1c4199987ebb-merged.mount: Deactivated successfully. Nov 27 04:58:16 localhost nova_compute[284026]: 2025-11-27 09:58:16.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:16 localhost systemd[1]: tmp-crun.rOJJGK.mount: Deactivated successfully. Nov 27 04:58:16 localhost podman[305854]: 2025-11-27 09:58:16.831776226 +0000 UTC m=+0.099149412 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, release=553, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., architecture=x86_64, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, io.buildah.version=1.33.12, name=rhceph, vcs-type=git, maintainer=Guillaume Abrioux , GIT_BRANCH=main) Nov 27 04:58:16 localhost podman[305854]: 2025-11-27 09:58:16.963359406 +0000 UTC m=+0.230732602 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, name=rhceph, vendor=Red Hat, Inc., GIT_CLEAN=True, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, vcs-type=git, architecture=x86_64, release=553, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:58:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:58:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:58:17 localhost podman[305906]: 2025-11-27 09:58:17.302821442 +0000 UTC m=+0.095919304 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:58:17 localhost podman[305906]: 2025-11-27 09:58:17.344073969 +0000 UTC m=+0.137171851 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:58:17 localhost podman[305907]: 2025-11-27 09:58:17.358591958 +0000 UTC m=+0.148166866 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 04:58:17 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:58:17 localhost podman[305907]: 2025-11-27 09:58:17.393050743 +0000 UTC m=+0.182625631 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:58:17 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:58:17 localhost nova_compute[284026]: 2025-11-27 09:58:17.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.098 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.763 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:58:18 localhost nova_compute[284026]: 2025-11-27 09:58:18.763 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.228 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.465s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.294 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.295 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing) e16 sync_obtain_latest_monmap Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing) e16 sync_obtain_latest_monmap obtained monmap e16 Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.518 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.520 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11492MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.520 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.521 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.592 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.593 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.593 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:58:19 localhost nova_compute[284026]: 2025-11-27 09:58:19.657 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).mds e17 new map Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).mds e17 print_map#012e17#012enable_multiple, ever_enabled_multiple: 1,1#012default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,12=quiesce subvolumes}#012legacy client fscid: 1#012 #012Filesystem 'cephfs' (1)#012fs_name#011cephfs#012epoch#01116#012flags#01112 joinable allow_snaps allow_multimds_snaps#012created#0112025-11-27T08:07:00.938837+0000#012modified#0112025-11-27T09:51:38.628845+0000#012tableserver#0110#012root#0110#012session_timeout#01160#012session_autoclose#011300#012max_file_size#0111099511627776#012required_client_features#011{}#012last_failure#0110#012last_failure_osd_epoch#01181#012compat#011compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2,12=quiesce subvolumes}#012max_mds#0111#012in#0110#012up#011{0=26518}#012failed#011#012damaged#011#012stopped#011#012data_pools#011[6]#012metadata_pool#0117#012inline_data#011disabled#012balancer#011#012bal_rank_mask#011-1#012standby_count_wanted#0111#012qdb_cluster#011leader: 26518 members: 26518#012[mds.mds.np0005537445.rkchqo{0:26518} state up:active seq 13 addr [v2:172.18.0.107:6808/3417254461,v1:172.18.0.107:6809/3417254461] compat {c=[1],r=[1],i=[17ff]}]#012 #012 #012Standby daemons:#012 #012[mds.mds.np0005537446.hkzfou{-1:17040} state up:standby seq 1 addr [v2:172.18.0.108:6808/2569123203,v1:172.18.0.108:6809/2569123203] compat {c=[1],r=[1],i=[17ff]}]#012[mds.mds.np0005537444.iqtbiy{-1:17052} state up:standby seq 1 addr [v2:172.18.0.106:6808/4180494082,v1:172.18.0.106:6809/4180494082] compat {c=[1],r=[1],i=[17ff]}] Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).osd e87 crush map has features 3314933000854323200, adjusting msgr requires Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).osd e87 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).osd e87 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).osd e87 crush map has features 432629239337189376, adjusting msgr requires Nov 27 04:58:19 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:19 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Deploying daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring osd.1 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring osd.0 (monmap changed)... Nov 27 04:58:19 localhost ceph-mon[305520]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:19 localhost ceph-mon[305520]: mon.np0005537446@-1(synchronizing).paxosservice(auth 1..41) refresh upgraded, format 0 -> 3 Nov 27 04:58:19 localhost ceph-mgr[290377]: ms_deliver_dispatch: unhandled message 0x561963d90000 mon_map magic: 0 from mon.0 v2:172.18.0.103:3300/0 Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.018 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.124 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.467s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.130 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.157 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.160 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:58:20 localhost nova_compute[284026]: 2025-11-27 09:58:20.160 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.640s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:58:21 localhost ceph-mon[305520]: mon.np0005537446@-1(probing) e17 my rank is now 2 (was -1) Nov 27 04:58:21 localhost ceph-mon[305520]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:58:21 localhost ceph-mon[305520]: paxos.2).electionLogic(0) init, first boot, initializing epoch at 1 Nov 27 04:58:21 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.161 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.162 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.162 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.918 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.918 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.919 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:58:22 localhost nova_compute[284026]: 2025-11-27 09:58:22.919 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:58:23 localhost nova_compute[284026]: 2025-11-27 09:58:23.130 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:23 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:58:23 localhost nova_compute[284026]: 2025-11-27 09:58:23.493 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:58:23 localhost nova_compute[284026]: 2025-11-27 09:58:23.508 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:58:23 localhost nova_compute[284026]: 2025-11-27 09:58:23.508 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:58:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:58:23 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:58:23 localhost podman[306070]: 2025-11-27 09:58:23.995243401 +0000 UTC m=+0.087166500 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, vcs-type=git, build-date=2025-08-20T13:12:41, maintainer=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, architecture=x86_64, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., distribution-scope=public, managed_by=edpm_ansible) Nov 27 04:58:24 localhost podman[306070]: 2025-11-27 09:58:24.036995911 +0000 UTC m=+0.128918960 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, io.openshift.expose-services=, name=ubi9-minimal, build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, io.buildah.version=1.33.7, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc., io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, distribution-scope=public, release=1755695350) Nov 27 04:58:24 localhost systemd[1]: tmp-crun.qXLXmF.mount: Deactivated successfully. Nov 27 04:58:24 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:58:24 localhost podman[306069]: 2025-11-27 09:58:24.059033352 +0000 UTC m=+0.152873632 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:58:24 localhost podman[306069]: 2025-11-27 09:58:24.076008137 +0000 UTC m=+0.169848407 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3) Nov 27 04:58:24 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:58:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:58:24 localhost podman[306108]: 2025-11-27 09:58:24.983912903 +0000 UTC m=+0.081491557 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:58:24 localhost systemd[1]: tmp-crun.Szod3d.mount: Deactivated successfully. Nov 27 04:58:25 localhost podman[306108]: 2025-11-27 09:58:25.027070371 +0000 UTC m=+0.124649045 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 04:58:25 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:58:25 localhost nova_compute[284026]: 2025-11-27 09:58:25.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:25 localhost openstack_network_exporter[244641]: ERROR 09:58:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:58:25 localhost openstack_network_exporter[244641]: ERROR 09:58:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:58:25 localhost openstack_network_exporter[244641]: ERROR 09:58:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:58:25 localhost openstack_network_exporter[244641]: ERROR 09:58:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:58:25 localhost openstack_network_exporter[244641]: Nov 27 04:58:25 localhost openstack_network_exporter[244641]: ERROR 09:58:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:58:25 localhost openstack_network_exporter[244641]: Nov 27 04:58:28 localhost nova_compute[284026]: 2025-11-27 09:58:28.180 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:28 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 handle_auth_request failed to assign global_id Nov 27 04:58:28 localhost systemd-logind[761]: Session 69 logged out. Waiting for processes to exit. Nov 27 04:58:28 localhost systemd[1]: session-69.scope: Deactivated successfully. Nov 27 04:58:28 localhost systemd[1]: session-69.scope: Consumed 31.599s CPU time. Nov 27 04:58:28 localhost systemd-logind[761]: Removed session 69. Nov 27 04:58:28 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 handle_auth_request failed to assign global_id Nov 27 04:58:28 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 handle_auth_request failed to assign global_id Nov 27 04:58:28 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 handle_auth_request failed to assign global_id Nov 27 04:58:28 localhost sshd[306468]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:58:28 localhost systemd-logind[761]: New session 72 of user ceph-admin. Nov 27 04:58:28 localhost systemd[1]: Started Session 72 of User ceph-admin. Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #13. Immutable memtables: 0. Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.505540) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 3] Flushing memtable with next log file: 13 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509505721, "job": 3, "event": "flush_started", "num_memtables": 1, "num_entries": 12153, "num_deletes": 257, "total_data_size": 22841890, "memory_usage": 23722056, "flush_reason": "Manual Compaction"} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 3] Level-0 flush table #14: started Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(synchronizing).osd e87 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(synchronizing).osd e87 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(synchronizing).osd e88 e88: 6 total, 6 up, 6 in Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537445 calling monitor election Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537444 calling monitor election Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537444 is new leader, mons np0005537444,np0005537445 in quorum (ranks 0,1) Nov 27 04:58:29 localhost ceph-mon[305520]: Health check failed: 1/3 mons down, quorum np0005537444,np0005537445 (MON_DOWN) Nov 27 04:58:29 localhost ceph-mon[305520]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm; 1/3 mons down, quorum np0005537444,np0005537445 Nov 27 04:58:29 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:58:29 localhost ceph-mon[305520]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:58:29 localhost ceph-mon[305520]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:58:29 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:58:29 localhost ceph-mon[305520]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:58:29 localhost ceph-mon[305520]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:58:29 localhost ceph-mon[305520]: [WRN] MON_DOWN: 1/3 mons down, quorum np0005537444,np0005537445 Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446 (rank 2) addr [v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0] is down (out of quorum) Nov 27 04:58:29 localhost ceph-mon[305520]: Reconfig service osd.default_drive_group Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' Nov 27 04:58:29 localhost ceph-mon[305520]: from='client.? 172.18.0.200:0/601215284' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26831 172.18.0.106:0/3128380996' entity='mgr.np0005537444.utkloe' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='client.? ' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: Activating manager daemon np0005537445.wmiblp Nov 27 04:58:29 localhost ceph-mon[305520]: from='client.? ' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:58:29 localhost ceph-mon[305520]: Manager daemon np0005537445.wmiblp is now available Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"}]': finished Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"} : dispatch Nov 27 04:58:29 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.np0005537443.localdomain.devices.0"}]': finished Nov 27 04:58:29 localhost ceph-mon[305520]: log_channel(cluster) log [INF] : mon.np0005537446 calling monitor election Nov 27 04:58:29 localhost ceph-mon[305520]: paxos.2).electionLogic(75) init, last seen epoch 75, mid-election, bumping Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(electing) e17 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 _apply_compatset_features enabling new quorum features: compat={},rocompat={},incompat={4=support erasure code pools,5=new-style osdmap encoding,6=support isa/lrc erasure code,7=support shec erasure code} Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 _apply_compatset_features enabling new quorum features: compat={},rocompat={},incompat={8=support monmap features,9=luminous ondisk layout,10=mimic ondisk layout,11=nautilus ondisk layout,12=octopus ondisk layout,13=pacific ondisk layout,14=quincy ondisk layout,15=reef ondisk layout} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509590212, "cf_name": "default", "job": 3, "event": "table_file_creation", "file_number": 14, "file_size": 18167384, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 12158, "table_properties": {"data_size": 18099569, "index_size": 37181, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 29573, "raw_key_size": 315356, "raw_average_key_size": 26, "raw_value_size": 17897836, "raw_average_value_size": 1514, "num_data_blocks": 1411, "num_entries": 11817, "num_filter_entries": 11817, "num_deletions": 256, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237499, "oldest_key_time": 1764237499, "file_creation_time": 1764237509, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 14, "seqno_to_time_mapping": "N/A"}} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 3] Flush lasted 84730 microseconds, and 39231 cpu microseconds. Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.590293) [db/flush_job.cc:967] [default] [JOB 3] Level-0 flush table #14: 18167384 bytes OK Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.590322) [db/memtable_list.cc:519] [default] Level-0 commit table #14 started Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.594495) [db/memtable_list.cc:722] [default] Level-0 commit table #14: memtable #1 done Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.594558) EVENT_LOG_v1 {"time_micros": 1764237509594515, "job": 3, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [2, 0, 0, 0, 0, 0, 0], "immutable_memtables": 0} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.594579) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: files[2 0 0 0 0 0 0] max score 0.50 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 3] Try to delete WAL files size 22756946, prev total WAL file size 22759107, number of live WAL files 2. Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.598799) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003131303434' seq:72057594037927935, type:22 .. '7061786F73003131323936' seq:0, type:0; will stop at (end) Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 4] Compacting 2@0 files to L6, score -1.00 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 3 Base level 0, inputs: [14(17MB) 8(1762B)] Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509598923, "job": 4, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [14, 8], "score": -1, "input_data_size": 18169146, "oldest_snapshot_seqno": -1} Nov 27 04:58:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 collect_metadata vda: no unique device id for vda: fallback method has no model nor serial Nov 27 04:58:29 localhost ceph-mon[305520]: mgrc update_daemon_metadata mon.np0005537446 metadata {addrs=[v2:172.18.0.105:3300/0,v1:172.18.0.105:6789/0],arch=x86_64,ceph_release=reef,ceph_version=ceph version 18.2.1-361.el9cp (439dcd6094d413840eb2ec590fe2194ec616687f) reef (stable),ceph_version_short=18.2.1-361.el9cp,compression_algorithms=none, snappy, zlib, zstd, lz4,container_hostname=np0005537446.localdomain,container_image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest,cpu=AMD EPYC-Rome Processor,device_ids=,device_paths=vda=/dev/disk/by-path/pci-0000:00:04.0,devices=vda,distro=rhel,distro_description=Red Hat Enterprise Linux 9.6 (Plow),distro_version=9.6,hostname=np0005537446.localdomain,kernel_description=#1 SMP PREEMPT_DYNAMIC Wed Apr 12 10:45:03 EDT 2023,kernel_version=5.14.0-284.11.1.el9_2.x86_64,mem_swap_kb=1048572,mem_total_kb=16116612,os=Linux} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 4] Generated table #15: 11566 keys, 18163960 bytes, temperature: kUnknown Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509690908, "cf_name": "default", "job": 4, "event": "table_file_creation", "file_number": 15, "file_size": 18163960, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 18096803, "index_size": 37168, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 28933, "raw_key_size": 310500, "raw_average_key_size": 26, "raw_value_size": 17898362, "raw_average_value_size": 1547, "num_data_blocks": 1411, "num_entries": 11566, "num_filter_entries": 11566, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237509, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 15, "seqno_to_time_mapping": "N/A"}} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.691282) [db/compaction/compaction_job.cc:1663] [default] [JOB 4] Compacted 2@0 files to L6 => 18163960 bytes Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.693168) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 197.2 rd, 197.2 wr, level 6, files in(2, 0) out(1 +0 blob) MB in(17.3, 0.0 +0.0 blob) out(17.3 +0.0 blob), read-write-amplify(2.0) write-amplify(1.0) OK, records in: 11822, records dropped: 256 output_compression: NoCompression Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.693198) EVENT_LOG_v1 {"time_micros": 1764237509693185, "job": 4, "event": "compaction_finished", "compaction_time_micros": 92113, "compaction_time_cpu_micros": 46444, "output_level": 6, "num_output_files": 1, "total_output_size": 18163960, "num_input_records": 11822, "num_output_records": 11566, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000014.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509695644, "job": 4, "event": "table_file_deletion", "file_number": 14} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000008.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237509695724, "job": 4, "event": "table_file_deletion", "file_number": 8} Nov 27 04:58:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:58:29.598613) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:58:30 localhost nova_compute[284026]: 2025-11-27 09:58:30.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:30 localhost podman[306579]: 2025-11-27 09:58:30.104683899 +0000 UTC m=+0.113712772 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, vcs-type=git, io.openshift.expose-services=, name=rhceph, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, RELEASE=main, ceph=True, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git) Nov 27 04:58:30 localhost podman[306579]: 2025-11-27 09:58:30.244840559 +0000 UTC m=+0.253869402 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, RELEASE=main, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, release=553, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, com.redhat.component=rhceph-container, io.openshift.expose-services=, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, architecture=x86_64, GIT_CLEAN=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, build-date=2025-09-24T08:57:55) Nov 27 04:58:30 localhost ceph-mon[305520]: mon.np0005537446 calling monitor election Nov 27 04:58:30 localhost ceph-mon[305520]: mon.np0005537446 calling monitor election Nov 27 04:58:30 localhost ceph-mon[305520]: mon.np0005537445 calling monitor election Nov 27 04:58:30 localhost ceph-mon[305520]: mon.np0005537444 calling monitor election Nov 27 04:58:30 localhost ceph-mon[305520]: mon.np0005537444 is new leader, mons np0005537444,np0005537445,np0005537446 in quorum (ranks 0,1,2) Nov 27 04:58:30 localhost ceph-mon[305520]: Health check cleared: MON_DOWN (was: 1/3 mons down, quorum np0005537444,np0005537445) Nov 27 04:58:30 localhost ceph-mon[305520]: Health detail: HEALTH_WARN 2 stray daemon(s) not managed by cephadm; 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:58:30 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_DAEMON: 2 stray daemon(s) not managed by cephadm Nov 27 04:58:30 localhost ceph-mon[305520]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 04:58:30 localhost ceph-mon[305520]: stray daemon mgr.np0005537442.byetac on host np0005537442.localdomain not managed by cephadm Nov 27 04:58:30 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_HOST: 2 stray host(s) with 2 daemon(s) not managed by cephadm Nov 27 04:58:30 localhost ceph-mon[305520]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 04:58:30 localhost ceph-mon[305520]: stray host np0005537442.localdomain has 1 stray daemons: ['mgr.np0005537442.byetac'] Nov 27 04:58:30 localhost ceph-mon[305520]: [27/Nov/2025:09:58:29] ENGINE Bus STARTING Nov 27 04:58:30 localhost ceph-mon[305520]: [27/Nov/2025:09:58:29] ENGINE Serving on http://172.18.0.107:8765 Nov 27 04:58:31 localhost ceph-mon[305520]: [27/Nov/2025:09:58:29] ENGINE Serving on https://172.18.0.107:7150 Nov 27 04:58:31 localhost ceph-mon[305520]: [27/Nov/2025:09:58:29] ENGINE Bus STARTED Nov 27 04:58:31 localhost ceph-mon[305520]: [27/Nov/2025:09:58:29] ENGINE Client ('172.18.0.107', 52852) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:58:31 localhost ceph-mon[305520]: Health check cleared: CEPHADM_STRAY_DAEMON (was: 2 stray daemon(s) not managed by cephadm) Nov 27 04:58:31 localhost ceph-mon[305520]: Health check cleared: CEPHADM_STRAY_HOST (was: 2 stray host(s) with 2 daemon(s) not managed by cephadm) Nov 27 04:58:31 localhost ceph-mon[305520]: Cluster is now healthy Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1019560329 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:33 localhost nova_compute[284026]: 2025-11-27 09:58:33.218 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:58:33 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:58:33 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:33 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:33 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:58:33 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:58:34 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:34 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:34 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:58:34 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:58:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:58:34 localhost podman[307217]: 2025-11-27 09:58:34.429154103 +0000 UTC m=+0.094354893 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 04:58:34 localhost podman[307217]: 2025-11-27 09:58:34.436733188 +0000 UTC m=+0.101933968 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 04:58:34 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:58:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:58:34 localhost podman[307292]: 2025-11-27 09:58:34.640396816 +0000 UTC m=+0.085438264 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 04:58:34 localhost podman[307292]: 2025-11-27 09:58:34.735056487 +0000 UTC m=+0.180098005 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:58:34 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:58:35 localhost nova_compute[284026]: 2025-11-27 09:58:35.049 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:35 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:58:35 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:58:35 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:58:35 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:58:35 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:36 localhost ceph-mon[305520]: Reconfiguring crash.np0005537444 (monmap changed)... Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:36 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537444 on np0005537444.localdomain Nov 27 04:58:36 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537444.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020045969 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:37 localhost ceph-mon[305520]: Health check failed: 2 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:58:37 localhost ceph-mon[305520]: Health check failed: 2 stray host(s) with 2 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:58:37 localhost ceph-mon[305520]: Reconfiguring osd.2 (monmap changed)... Nov 27 04:58:37 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch Nov 27 04:58:37 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:37 localhost ceph-mon[305520]: Reconfiguring daemon osd.2 on np0005537444.localdomain Nov 27 04:58:37 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:38 localhost nova_compute[284026]: 2025-11-27 09:58:38.260 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:38 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:38 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:38 localhost ceph-mon[305520]: Reconfiguring osd.5 (monmap changed)... Nov 27 04:58:38 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch Nov 27 04:58:38 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:38 localhost ceph-mon[305520]: Reconfiguring daemon osd.5 on np0005537444.localdomain Nov 27 04:58:38 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:38 localhost podman[242678]: time="2025-11-27T09:58:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:58:38 localhost podman[242678]: @ - - [27/Nov/2025:09:58:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:58:38 localhost podman[242678]: @ - - [27/Nov/2025:09:58:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19701 "" "Go-http-client/1.1" Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: Reconfiguring mds.mds.np0005537444.iqtbiy (monmap changed)... Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: Reconfiguring daemon mds.mds.np0005537444.iqtbiy on np0005537444.localdomain Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537444.iqtbiy", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:39 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537444.utkloe", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:40 localhost nova_compute[284026]: 2025-11-27 09:58:40.056 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:40 localhost ceph-mon[305520]: Reconfiguring mgr.np0005537444.utkloe (monmap changed)... Nov 27 04:58:40 localhost ceph-mon[305520]: Reconfiguring daemon mgr.np0005537444.utkloe on np0005537444.localdomain Nov 27 04:58:40 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:40 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:40 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:40 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537445.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:41 localhost ceph-mon[305520]: Reconfiguring crash.np0005537445 (monmap changed)... Nov 27 04:58:41 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537445 on np0005537445.localdomain Nov 27 04:58:41 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:41 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch Nov 27 04:58:41 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054576 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:42 localhost ceph-mon[305520]: Reconfiguring osd.1 (monmap changed)... Nov 27 04:58:42 localhost ceph-mon[305520]: Reconfiguring daemon osd.1 on np0005537445.localdomain Nov 27 04:58:42 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:42 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:42 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch Nov 27 04:58:42 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:42 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost nova_compute[284026]: 2025-11-27 09:58:43.263 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:58:43.568 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:58:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:58:43.569 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:58:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:58:43.570 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:58:43 localhost ceph-mon[305520]: Reconfiguring osd.4 (monmap changed)... Nov 27 04:58:43 localhost ceph-mon[305520]: Reconfiguring daemon osd.4 on np0005537445.localdomain Nov 27 04:58:43 localhost ceph-mon[305520]: Saving service mon spec with placement label:mon Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:43 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537445.rkchqo", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:44 localhost ceph-mon[305520]: Reconfiguring mds.mds.np0005537445.rkchqo (monmap changed)... Nov 27 04:58:44 localhost ceph-mon[305520]: Reconfiguring daemon mds.mds.np0005537445.rkchqo on np0005537445.localdomain Nov 27 04:58:44 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:44 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:44 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:44 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537445.wmiblp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:45 localhost nova_compute[284026]: 2025-11-27 09:58:45.062 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:45 localhost ceph-mon[305520]: Reconfiguring mgr.np0005537445.wmiblp (monmap changed)... Nov 27 04:58:45 localhost ceph-mon[305520]: Reconfiguring daemon mgr.np0005537445.wmiblp on np0005537445.localdomain Nov 27 04:58:45 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:45 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:58:45 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:46 localhost podman[307583]: Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.573502802 +0000 UTC m=+0.080300795 container create 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, GIT_CLEAN=True, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, vcs-type=git, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, CEPH_POINT_RELEASE=, name=rhceph, com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, description=Red Hat Ceph Storage 7, RELEASE=main) Nov 27 04:58:46 localhost systemd[1]: Started libpod-conmon-527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43.scope. Nov 27 04:58:46 localhost systemd[1]: Started libcrun container. Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.540898183 +0000 UTC m=+0.047696226 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.643756515 +0000 UTC m=+0.150554498 container init 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., version=7, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, vcs-type=git, build-date=2025-09-24T08:57:55, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, ceph=True, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, architecture=x86_64) Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.654880825 +0000 UTC m=+0.161678818 container start 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, RELEASE=main, maintainer=Guillaume Abrioux , vendor=Red Hat, Inc., release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_BRANCH=main, name=rhceph, GIT_CLEAN=True, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, version=7, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.655159882 +0000 UTC m=+0.161957905 container attach 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, io.openshift.expose-services=, release=553, CEPH_POINT_RELEASE=, RELEASE=main, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_CLEAN=True, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, name=rhceph, com.redhat.component=rhceph-container, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, version=7, vcs-type=git) Nov 27 04:58:46 localhost sad_knuth[307598]: 167 167 Nov 27 04:58:46 localhost systemd[1]: libpod-527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43.scope: Deactivated successfully. Nov 27 04:58:46 localhost podman[307583]: 2025-11-27 09:58:46.66065792 +0000 UTC m=+0.167455933 container died 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, com.redhat.license_terms=https://www.redhat.com/agreements, name=rhceph, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, release=553, ceph=True, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_CLEAN=True, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, build-date=2025-09-24T08:57:55, GIT_REPO=https://github.com/ceph/ceph-container.git, architecture=x86_64, distribution-scope=public, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux ) Nov 27 04:58:46 localhost podman[307603]: 2025-11-27 09:58:46.757204521 +0000 UTC m=+0.087603011 container remove 527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=sad_knuth, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, ceph=True, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, io.openshift.expose-services=, architecture=x86_64, description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, RELEASE=main, version=7, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vendor=Red Hat, Inc., vcs-type=git, io.openshift.tags=rhceph ceph) Nov 27 04:58:46 localhost systemd[1]: libpod-conmon-527ff8c9624ef224981fabe3b34359507abc0616b57509776d18917086e9fd43.scope: Deactivated successfully. Nov 27 04:58:46 localhost ceph-mon[305520]: Reconfiguring mon.np0005537445 (monmap changed)... Nov 27 04:58:46 localhost ceph-mon[305520]: Reconfiguring daemon mon.np0005537445 on np0005537445.localdomain Nov 27 04:58:46 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:46 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:46 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:46 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "client.crash.np0005537446.localdomain", "caps": ["mon", "profile crash", "mgr", "profile crash"]} : dispatch Nov 27 04:58:46 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054729 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:58:47 localhost podman[307669]: 2025-11-27 09:58:47.478622312 +0000 UTC m=+0.086224575 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, managed_by=edpm_ansible, tcib_managed=true, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 04:58:47 localhost podman[307677]: Nov 27 04:58:47 localhost podman[307669]: 2025-11-27 09:58:47.497873831 +0000 UTC m=+0.105476104 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:58:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:58:47 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.551556857 +0000 UTC m=+0.137218319 container create 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, GIT_CLEAN=True, description=Red Hat Ceph Storage 7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, GIT_BRANCH=main, vcs-type=git, io.buildah.version=1.33.12, ceph=True, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, name=rhceph, release=553) Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.465355974 +0000 UTC m=+0.051017456 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:47 localhost systemd[1]: var-lib-containers-storage-overlay-478976a43b87a3f3975b7041b3b474b4def62475cb0d8da583ee025a9e86da0c-merged.mount: Deactivated successfully. Nov 27 04:58:47 localhost systemd[1]: Started libpod-conmon-29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1.scope. Nov 27 04:58:47 localhost systemd[1]: Started libcrun container. Nov 27 04:58:47 localhost podman[307704]: 2025-11-27 09:58:47.632623102 +0000 UTC m=+0.099201445 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, config_id=ovn_metadata_agent) Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.639729402 +0000 UTC m=+0.225390864 container init 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , distribution-scope=public, description=Red Hat Ceph Storage 7, vcs-type=git, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., release=553, GIT_CLEAN=True, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, RELEASE=main, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, name=rhceph, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.650476832 +0000 UTC m=+0.236138294 container start 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, io.openshift.tags=rhceph ceph, distribution-scope=public, name=rhceph, release=553, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, version=7, ceph=True, vcs-type=git, io.k8s.description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, build-date=2025-09-24T08:57:55, GIT_BRANCH=main, maintainer=Guillaume Abrioux , description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, RELEASE=main, vendor=Red Hat, Inc.) Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.650830352 +0000 UTC m=+0.236491854 container attach 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, io.openshift.expose-services=, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.description=Red Hat Ceph Storage 7, version=7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, name=rhceph, maintainer=Guillaume Abrioux , io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, RELEASE=main, architecture=x86_64, build-date=2025-09-24T08:57:55, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:58:47 localhost silly_visvesvaraya[307718]: 167 167 Nov 27 04:58:47 localhost systemd[1]: libpod-29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1.scope: Deactivated successfully. Nov 27 04:58:47 localhost podman[307677]: 2025-11-27 09:58:47.65336017 +0000 UTC m=+0.239021632 container died 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, release=553, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, version=7, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, com.redhat.component=rhceph-container, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:58:47 localhost podman[307704]: 2025-11-27 09:58:47.668029285 +0000 UTC m=+0.134607578 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true, config_id=ovn_metadata_agent) Nov 27 04:58:47 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:58:47 localhost podman[307728]: 2025-11-27 09:58:47.750167819 +0000 UTC m=+0.083686806 container remove 29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=silly_visvesvaraya, release=553, maintainer=Guillaume Abrioux , GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, RELEASE=main, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, vcs-type=git, GIT_CLEAN=True, com.redhat.component=rhceph-container, name=rhceph, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, architecture=x86_64, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553) Nov 27 04:58:47 localhost systemd[1]: libpod-conmon-29f121cffb0905264c2e80099050087203ed19f09de46ab5eef78dfe0191adf1.scope: Deactivated successfully. Nov 27 04:58:47 localhost ceph-mon[305520]: Reconfiguring crash.np0005537446 (monmap changed)... Nov 27 04:58:47 localhost ceph-mon[305520]: Reconfiguring daemon crash.np0005537446 on np0005537446.localdomain Nov 27 04:58:47 localhost ceph-mon[305520]: Reconfiguring osd.0 (monmap changed)... Nov 27 04:58:47 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch Nov 27 04:58:47 localhost ceph-mon[305520]: Reconfiguring daemon osd.0 on np0005537446.localdomain Nov 27 04:58:47 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:48 localhost nova_compute[284026]: 2025-11-27 09:58:48.302 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:48 localhost systemd[1]: var-lib-containers-storage-overlay-9b5a5edfdf52b68f5991b1bbcf2cbce1d6bcfb352d5b87556464b27d314a21ee-merged.mount: Deactivated successfully. Nov 27 04:58:48 localhost podman[307805]: Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.622430123 +0000 UTC m=+0.087165250 container create b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, GIT_CLEAN=True, RELEASE=main, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, build-date=2025-09-24T08:57:55, name=rhceph, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, vcs-type=git, io.openshift.tags=rhceph ceph, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, CEPH_POINT_RELEASE=, version=7, vendor=Red Hat, Inc.) Nov 27 04:58:48 localhost systemd[1]: Started libpod-conmon-b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24.scope. Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.588689464 +0000 UTC m=+0.053424601 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:48 localhost systemd[1]: Started libcrun container. Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.710005843 +0000 UTC m=+0.174740970 container init b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, name=rhceph, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , io.buildah.version=1.33.12, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, GIT_CLEAN=True, RELEASE=main, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., architecture=x86_64, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3) Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.723338002 +0000 UTC m=+0.188073129 container start b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, name=rhceph, io.buildah.version=1.33.12, RELEASE=main, maintainer=Guillaume Abrioux , release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, vcs-type=git, GIT_BRANCH=main, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, com.redhat.component=rhceph-container, architecture=x86_64, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7) Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.723818775 +0000 UTC m=+0.188553902 container attach b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, name=rhceph, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, architecture=x86_64, io.openshift.expose-services=, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, distribution-scope=public, version=7, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, ceph=True, maintainer=Guillaume Abrioux ) Nov 27 04:58:48 localhost wizardly_moore[307820]: 167 167 Nov 27 04:58:48 localhost systemd[1]: libpod-b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24.scope: Deactivated successfully. Nov 27 04:58:48 localhost podman[307805]: 2025-11-27 09:58:48.728947253 +0000 UTC m=+0.193682390 container died b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, vcs-type=git, description=Red Hat Ceph Storage 7, distribution-scope=public, com.redhat.component=rhceph-container, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, CEPH_POINT_RELEASE=, ceph=True, maintainer=Guillaume Abrioux , vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, GIT_BRANCH=main, architecture=x86_64, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.expose-services=, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.buildah.version=1.33.12, RELEASE=main, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, GIT_CLEAN=True) Nov 27 04:58:48 localhost podman[307825]: 2025-11-27 09:58:48.845581756 +0000 UTC m=+0.102026990 container remove b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=wizardly_moore, io.buildah.version=1.33.12, vendor=Red Hat, Inc., ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, version=7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, release=553, io.openshift.expose-services=, maintainer=Guillaume Abrioux , name=rhceph, architecture=x86_64, distribution-scope=public, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, RELEASE=main, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, GIT_CLEAN=True, vcs-type=git, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:58:48 localhost systemd[1]: libpod-conmon-b390691ce35b402c42839bf2ac42532f3819505c2fbc0367319d4adbe127fe24.scope: Deactivated successfully. Nov 27 04:58:48 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:48 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:48 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch Nov 27 04:58:48 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:48 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:49 localhost systemd[1]: var-lib-containers-storage-overlay-f9ade753389c9047ae92f3fa591e39ada77184db46e979b4540b5150edc0c7cb-merged.mount: Deactivated successfully. Nov 27 04:58:49 localhost podman[307902]: Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.774648381 +0000 UTC m=+0.086429810 container create ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, RELEASE=main, name=rhceph, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, vendor=Red Hat, Inc., distribution-scope=public, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553) Nov 27 04:58:49 localhost systemd[1]: Started libpod-conmon-ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9.scope. Nov 27 04:58:49 localhost systemd[1]: Started libcrun container. Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.739480754 +0000 UTC m=+0.051262203 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.849377875 +0000 UTC m=+0.161159304 container init ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, ceph=True, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, GIT_CLEAN=True, com.redhat.component=rhceph-container, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, vendor=Red Hat, Inc., release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, RELEASE=main, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, architecture=x86_64, io.openshift.expose-services=, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, name=rhceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.864183824 +0000 UTC m=+0.175965253 container start ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, vendor=Red Hat, Inc., io.buildah.version=1.33.12, name=rhceph, maintainer=Guillaume Abrioux , architecture=x86_64, release=553, com.redhat.component=rhceph-container, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., build-date=2025-09-24T08:57:55, ceph=True, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7) Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.86443409 +0000 UTC m=+0.176215559 container attach ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, version=7, name=rhceph, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/agreements, release=553, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, RELEASE=main, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, ceph=True, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.openshift.expose-services=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, distribution-scope=public, io.buildah.version=1.33.12, vcs-type=git, architecture=x86_64) Nov 27 04:58:49 localhost epic_gates[307919]: 167 167 Nov 27 04:58:49 localhost systemd[1]: libpod-ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9.scope: Deactivated successfully. Nov 27 04:58:49 localhost podman[307902]: 2025-11-27 09:58:49.86885859 +0000 UTC m=+0.180640069 container died ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, name=rhceph, com.redhat.component=rhceph-container, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, release=553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., distribution-scope=public, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, version=7, GIT_CLEAN=True, maintainer=Guillaume Abrioux , ceph=True, io.openshift.expose-services=, build-date=2025-09-24T08:57:55) Nov 27 04:58:49 localhost ceph-mon[305520]: Reconfiguring osd.3 (monmap changed)... Nov 27 04:58:49 localhost ceph-mon[305520]: Reconfiguring daemon osd.3 on np0005537446.localdomain Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:49 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mds.mds.np0005537446.hkzfou", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]} : dispatch Nov 27 04:58:49 localhost podman[307924]: 2025-11-27 09:58:49.971425133 +0000 UTC m=+0.091324871 container remove ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=epic_gates, io.openshift.expose-services=, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, io.buildah.version=1.33.12, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, com.redhat.component=rhceph-container, ceph=True, com.redhat.license_terms=https://www.redhat.com/agreements, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., maintainer=Guillaume Abrioux , GIT_BRANCH=main, GIT_CLEAN=True, vendor=Red Hat, Inc., description=Red Hat Ceph Storage 7, CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, vcs-type=git, io.openshift.tags=rhceph ceph) Nov 27 04:58:49 localhost systemd[1]: libpod-conmon-ed9a0f70493639ccb0f7bd00e6a07de8225f67900c34c59cf2affa7afc0fd7a9.scope: Deactivated successfully. Nov 27 04:58:50 localhost nova_compute[284026]: 2025-11-27 09:58:50.062 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:50 localhost systemd[1]: tmp-crun.RYPuv2.mount: Deactivated successfully. Nov 27 04:58:50 localhost systemd[1]: var-lib-containers-storage-overlay-d257db43add570764c2b9be9e6b6bb0c9294f715987b7fef72c1a9d219301744-merged.mount: Deactivated successfully. Nov 27 04:58:50 localhost podman[307994]: Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.759983053 +0000 UTC m=+0.087403947 container create 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_BRANCH=main, io.buildah.version=1.33.12, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, io.openshift.expose-services=, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, build-date=2025-09-24T08:57:55, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_CLEAN=True, release=553, maintainer=Guillaume Abrioux , architecture=x86_64, version=7, RELEASE=main, name=rhceph, vcs-type=git) Nov 27 04:58:50 localhost systemd[1]: Started libpod-conmon-7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452.scope. Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.725818232 +0000 UTC m=+0.053239176 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:50 localhost systemd[1]: Started libcrun container. Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.841854958 +0000 UTC m=+0.169275862 container init 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, RELEASE=main, GIT_BRANCH=main, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, release=553, architecture=x86_64, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, vcs-type=git, version=7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, CEPH_POINT_RELEASE=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux ) Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.852799473 +0000 UTC m=+0.180220367 container start 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, io.buildah.version=1.33.12, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.openshift.expose-services=, vcs-type=git, com.redhat.component=rhceph-container, release=553, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, version=7, architecture=x86_64, description=Red Hat Ceph Storage 7, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.tags=rhceph ceph, RELEASE=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.853126792 +0000 UTC m=+0.180547726 container attach 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, version=7, io.openshift.expose-services=, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, vcs-type=git, GIT_BRANCH=main, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, GIT_CLEAN=True, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., com.redhat.component=rhceph-container, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:58:50 localhost competent_greider[308010]: 167 167 Nov 27 04:58:50 localhost systemd[1]: libpod-7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452.scope: Deactivated successfully. Nov 27 04:58:50 localhost podman[307994]: 2025-11-27 09:58:50.856918135 +0000 UTC m=+0.184339069 container died 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, com.redhat.component=rhceph-container, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, GIT_BRANCH=main, ceph=True, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_REPO=https://github.com/ceph/ceph-container.git, release=553, architecture=x86_64, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, maintainer=Guillaume Abrioux , name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 04:58:50 localhost ceph-mon[305520]: Reconfiguring mds.mds.np0005537446.hkzfou (monmap changed)... Nov 27 04:58:50 localhost ceph-mon[305520]: Reconfiguring daemon mds.mds.np0005537446.hkzfou on np0005537446.localdomain Nov 27 04:58:50 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:50 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:50 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:50 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get-or-create", "entity": "mgr.np0005537446.nfulyg", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch Nov 27 04:58:50 localhost podman[308015]: 2025-11-27 09:58:50.964147764 +0000 UTC m=+0.096773599 container remove 7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=competent_greider, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-type=git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, distribution-scope=public, io.openshift.tags=rhceph ceph, name=rhceph, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, version=7, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_BRANCH=main, io.buildah.version=1.33.12, CEPH_POINT_RELEASE=, description=Red Hat Ceph Storage 7, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, build-date=2025-09-24T08:57:55, ceph=True, io.openshift.expose-services=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, architecture=x86_64, vendor=Red Hat, Inc.) Nov 27 04:58:50 localhost systemd[1]: libpod-conmon-7ff8d1ced38a275623c5878b117ae0fbdfafd671b070ad34fa7ad45d9c03c452.scope: Deactivated successfully. Nov 27 04:58:51 localhost systemd[1]: session-70.scope: Deactivated successfully. Nov 27 04:58:51 localhost systemd[1]: session-70.scope: Consumed 1.860s CPU time. Nov 27 04:58:51 localhost systemd-logind[761]: Session 70 logged out. Waiting for processes to exit. Nov 27 04:58:51 localhost systemd-logind[761]: Removed session 70. Nov 27 04:58:51 localhost systemd[1]: var-lib-containers-storage-overlay-cf968349f3d9432abac6aa08be6d6485426f331904faad69130af18df1acf6a9-merged.mount: Deactivated successfully. Nov 27 04:58:51 localhost podman[308085]: Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.809164404 +0000 UTC m=+0.084096847 container create a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, io.openshift.expose-services=, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, build-date=2025-09-24T08:57:55, ceph=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, name=rhceph, architecture=x86_64, com.redhat.component=rhceph-container, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, CEPH_POINT_RELEASE=, io.openshift.tags=rhceph ceph, release=553, version=7, GIT_BRANCH=main, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, description=Red Hat Ceph Storage 7) Nov 27 04:58:51 localhost systemd[1]: Started libpod-conmon-a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185.scope. Nov 27 04:58:51 localhost systemd[1]: Started libcrun container. Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.776371781 +0000 UTC m=+0.051304264 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.885562653 +0000 UTC m=+0.160495086 container init a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-type=git, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main, distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, ceph=True, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, name=rhceph, io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, GIT_CLEAN=True, vendor=Red Hat, Inc., com.redhat.component=rhceph-container) Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.894538945 +0000 UTC m=+0.169471378 container start a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, build-date=2025-09-24T08:57:55, distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., description=Red Hat Ceph Storage 7, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.openshift.tags=rhceph ceph, release=553, vcs-type=git, vendor=Red Hat, Inc., GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.component=rhceph-container, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, io.openshift.expose-services=, version=7, GIT_BRANCH=main, RELEASE=main, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.894896414 +0000 UTC m=+0.169828867 container attach a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, description=Red Hat Ceph Storage 7, RELEASE=main, build-date=2025-09-24T08:57:55, com.redhat.component=rhceph-container, GIT_BRANCH=main, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, release=553, CEPH_POINT_RELEASE=, maintainer=Guillaume Abrioux , architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, ceph=True, vcs-type=git, vendor=Red Hat, Inc., distribution-scope=public, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, version=7) Nov 27 04:58:51 localhost amazing_bartik[308101]: 167 167 Nov 27 04:58:51 localhost systemd[1]: libpod-a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185.scope: Deactivated successfully. Nov 27 04:58:51 localhost podman[308085]: 2025-11-27 09:58:51.899354375 +0000 UTC m=+0.174286838 container died a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, RELEASE=main, name=rhceph, vcs-type=git, CEPH_POINT_RELEASE=, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, version=7, ceph=True, io.openshift.expose-services=, release=553, maintainer=Guillaume Abrioux , io.k8s.description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph) Nov 27 04:58:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:51 localhost ceph-mon[305520]: Reconfiguring mgr.np0005537446.nfulyg (monmap changed)... Nov 27 04:58:51 localhost ceph-mon[305520]: Reconfiguring daemon mgr.np0005537446.nfulyg on np0005537446.localdomain Nov 27 04:58:51 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:51 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:58:51 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:52 localhost podman[308106]: 2025-11-27 09:58:52.00979184 +0000 UTC m=+0.096487040 container remove a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=amazing_bartik, build-date=2025-09-24T08:57:55, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, RELEASE=main, ceph=True, io.openshift.tags=rhceph ceph, release=553, GIT_CLEAN=True, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/agreements, io.openshift.expose-services=, vcs-type=git, version=7, io.k8s.description=Red Hat Ceph Storage 7, distribution-scope=public, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, description=Red Hat Ceph Storage 7, name=rhceph, CEPH_POINT_RELEASE=, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., io.buildah.version=1.33.12, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container) Nov 27 04:58:52 localhost systemd[1]: libpod-conmon-a603d3d47a385f760cb8983dd1e742b262451468ef99fda750aa4a8dfeff9185.scope: Deactivated successfully. Nov 27 04:58:52 localhost systemd[1]: var-lib-containers-storage-overlay-f61d83d044ba6e092f7a8b0060604c0f645f89f722eb48fa67798c34e97c04c2-merged.mount: Deactivated successfully. Nov 27 04:58:52 localhost ceph-mon[305520]: Reconfiguring mon.np0005537446 (monmap changed)... Nov 27 04:58:52 localhost ceph-mon[305520]: Reconfiguring daemon mon.np0005537446 on np0005537446.localdomain Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:52 localhost ceph-mon[305520]: from='mgr.26840 172.18.0.107:0/1484228113' entity='mgr.np0005537445.wmiblp' cmd={"prefix": "auth get", "entity": "mon."} : dispatch Nov 27 04:58:53 localhost nova_compute[284026]: 2025-11-27 09:58:53.304 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:53 localhost ceph-mon[305520]: Reconfiguring mon.np0005537444 (monmap changed)... Nov 27 04:58:53 localhost ceph-mon[305520]: Reconfiguring daemon mon.np0005537444 on np0005537444.localdomain Nov 27 04:58:53 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:53 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:53 localhost ceph-mon[305520]: from='mgr.26840 ' entity='mgr.np0005537445.wmiblp' Nov 27 04:58:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:58:54 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:58:55 localhost podman[308141]: 2025-11-27 09:58:54.999934594 +0000 UTC m=+0.086165372 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.expose-services=, maintainer=Red Hat, Inc., version=9.6, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, container_name=openstack_network_exporter, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, release=1755695350, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, vcs-type=git) Nov 27 04:58:55 localhost systemd[1]: tmp-crun.o6ErKs.mount: Deactivated successfully. Nov 27 04:58:55 localhost podman[308141]: 2025-11-27 09:58:55.041675749 +0000 UTC m=+0.127906527 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, maintainer=Red Hat, Inc., architecture=x86_64, container_name=openstack_network_exporter, managed_by=edpm_ansible, io.openshift.expose-services=, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, vcs-type=git, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.tags=minimal rhel9, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm) Nov 27 04:58:55 localhost nova_compute[284026]: 2025-11-27 09:58:55.085 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:58:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:58:55 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:58:55 localhost podman[308140]: 2025-11-27 09:58:55.043773865 +0000 UTC m=+0.134162006 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:58:55 localhost podman[308140]: 2025-11-27 09:58:55.128353304 +0000 UTC m=+0.218741415 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 04:58:55 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:58:55 localhost podman[308177]: 2025-11-27 09:58:55.190442977 +0000 UTC m=+0.087356605 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:58:55 localhost podman[308177]: 2025-11-27 09:58:55.202999616 +0000 UTC m=+0.099913244 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:58:55 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:58:55 localhost openstack_network_exporter[244641]: ERROR 09:58:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:58:55 localhost openstack_network_exporter[244641]: ERROR 09:58:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:58:55 localhost openstack_network_exporter[244641]: ERROR 09:58:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:58:55 localhost openstack_network_exporter[244641]: ERROR 09:58:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:58:55 localhost openstack_network_exporter[244641]: Nov 27 04:58:55 localhost openstack_network_exporter[244641]: ERROR 09:58:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:58:55 localhost openstack_network_exporter[244641]: Nov 27 04:58:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:58:58 localhost nova_compute[284026]: 2025-11-27 09:58:58.343 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:00 localhost nova_compute[284026]: 2025-11-27 09:59:00.089 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:01 localhost systemd[1]: Stopping User Manager for UID 1003... Nov 27 04:59:01 localhost systemd[303371]: Activating special unit Exit the Session... Nov 27 04:59:01 localhost systemd[303371]: Stopped target Main User Target. Nov 27 04:59:01 localhost systemd[303371]: Stopped target Basic System. Nov 27 04:59:01 localhost systemd[303371]: Stopped target Paths. Nov 27 04:59:01 localhost systemd[303371]: Stopped target Sockets. Nov 27 04:59:01 localhost systemd[303371]: Stopped target Timers. Nov 27 04:59:01 localhost systemd[303371]: Stopped Mark boot as successful after the user session has run 2 minutes. Nov 27 04:59:01 localhost systemd[303371]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 04:59:01 localhost systemd[303371]: Closed D-Bus User Message Bus Socket. Nov 27 04:59:01 localhost systemd[303371]: Stopped Create User's Volatile Files and Directories. Nov 27 04:59:01 localhost systemd[303371]: Removed slice User Application Slice. Nov 27 04:59:01 localhost systemd[303371]: Reached target Shutdown. Nov 27 04:59:01 localhost systemd[303371]: Finished Exit the Session. Nov 27 04:59:01 localhost systemd[303371]: Reached target Exit the Session. Nov 27 04:59:01 localhost systemd[1]: user@1003.service: Deactivated successfully. Nov 27 04:59:01 localhost systemd[1]: Stopped User Manager for UID 1003. Nov 27 04:59:01 localhost systemd[1]: Stopping User Runtime Directory /run/user/1003... Nov 27 04:59:01 localhost systemd[1]: run-user-1003.mount: Deactivated successfully. Nov 27 04:59:01 localhost systemd[1]: user-runtime-dir@1003.service: Deactivated successfully. Nov 27 04:59:01 localhost systemd[1]: Stopped User Runtime Directory /run/user/1003. Nov 27 04:59:01 localhost systemd[1]: Removed slice User Slice of UID 1003. Nov 27 04:59:01 localhost systemd[1]: user-1003.slice: Consumed 2.610s CPU time. Nov 27 04:59:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:03 localhost nova_compute[284026]: 2025-11-27 09:59:03.346 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:59:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:59:05 localhost systemd[1]: tmp-crun.L7o2X3.mount: Deactivated successfully. Nov 27 04:59:05 localhost podman[308203]: 2025-11-27 09:59:05.011262944 +0000 UTC m=+0.094211219 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:59:05 localhost podman[308202]: 2025-11-27 09:59:05.062632189 +0000 UTC m=+0.145589194 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:59:05 localhost podman[308203]: 2025-11-27 09:59:05.075043213 +0000 UTC m=+0.157991488 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:59:05 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:59:05 localhost nova_compute[284026]: 2025-11-27 09:59:05.093 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:05 localhost podman[308202]: 2025-11-27 09:59:05.109346337 +0000 UTC m=+0.192303322 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 04:59:05 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:59:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:08 localhost nova_compute[284026]: 2025-11-27 09:59:08.373 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:08 localhost podman[242678]: time="2025-11-27T09:59:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:59:08 localhost podman[242678]: @ - - [27/Nov/2025:09:59:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:59:08 localhost podman[242678]: @ - - [27/Nov/2025:09:59:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19708 "" "Go-http-client/1.1" Nov 27 04:59:10 localhost nova_compute[284026]: 2025-11-27 09:59:10.094 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:12 localhost nova_compute[284026]: 2025-11-27 09:59:12.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:13 localhost nova_compute[284026]: 2025-11-27 09:59:13.376 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:14 localhost nova_compute[284026]: 2025-11-27 09:59:14.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:14 localhost nova_compute[284026]: 2025-11-27 09:59:14.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:14 localhost nova_compute[284026]: 2025-11-27 09:59:14.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:15 localhost nova_compute[284026]: 2025-11-27 09:59:15.098 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:15 localhost nova_compute[284026]: 2025-11-27 09:59:15.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:15 localhost nova_compute[284026]: 2025-11-27 09:59:15.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 04:59:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:59:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:59:17 localhost podman[308251]: 2025-11-27 09:59:17.996405479 +0000 UTC m=+0.091202398 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 04:59:18 localhost podman[308251]: 2025-11-27 09:59:18.004923769 +0000 UTC m=+0.099720728 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, tcib_managed=true, container_name=ovn_metadata_agent, io.buildah.version=1.41.3) Nov 27 04:59:18 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:59:18 localhost podman[308252]: 2025-11-27 09:59:18.095901771 +0000 UTC m=+0.185932711 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, container_name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=multipathd, io.buildah.version=1.41.3) Nov 27 04:59:18 localhost podman[308252]: 2025-11-27 09:59:18.135483287 +0000 UTC m=+0.225514187 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=multipathd) Nov 27 04:59:18 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.413 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.729 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.748 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.749 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.767 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.768 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.769 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.769 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 04:59:18 localhost nova_compute[284026]: 2025-11-27 09:59:18.770 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:59:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:59:19 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3600785077' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.235 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.465s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.292 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.293 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.521 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.523 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11463MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.524 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.524 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.601 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.602 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.602 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 04:59:19 localhost nova_compute[284026]: 2025-11-27 09:59:19.666 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.099 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 04:59:20 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3215710815' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.144 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.477s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.151 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.175 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.177 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 04:59:20 localhost nova_compute[284026]: 2025-11-27 09:59:20.177 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.653s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:59:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.161 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.162 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.163 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.242 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.243 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.244 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 04:59:22 localhost nova_compute[284026]: 2025-11-27 09:59:22.245 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 04:59:23 localhost nova_compute[284026]: 2025-11-27 09:59:23.324 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 04:59:23 localhost nova_compute[284026]: 2025-11-27 09:59:23.337 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 04:59:23 localhost nova_compute[284026]: 2025-11-27 09:59:23.338 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 04:59:23 localhost nova_compute[284026]: 2025-11-27 09:59:23.417 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:25 localhost nova_compute[284026]: 2025-11-27 09:59:25.126 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:25 localhost openstack_network_exporter[244641]: ERROR 09:59:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:59:25 localhost openstack_network_exporter[244641]: ERROR 09:59:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:59:25 localhost openstack_network_exporter[244641]: ERROR 09:59:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:59:25 localhost openstack_network_exporter[244641]: ERROR 09:59:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:59:25 localhost openstack_network_exporter[244641]: Nov 27 04:59:25 localhost openstack_network_exporter[244641]: ERROR 09:59:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:59:25 localhost openstack_network_exporter[244641]: Nov 27 04:59:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:59:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:59:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:59:25 localhost podman[308333]: 2025-11-27 09:59:25.993220816 +0000 UTC m=+0.086846951 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:59:26 localhost podman[308333]: 2025-11-27 09:59:26.008372264 +0000 UTC m=+0.101998469 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 04:59:26 localhost podman[308335]: 2025-11-27 09:59:26.043810149 +0000 UTC m=+0.131092413 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, maintainer=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=edpm_ansible, vendor=Red Hat, Inc., io.openshift.expose-services=, release=1755695350, io.openshift.tags=minimal rhel9, name=ubi9-minimal, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-type=git, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6) Nov 27 04:59:26 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:59:26 localhost podman[308335]: 2025-11-27 09:59:26.08204421 +0000 UTC m=+0.169326484 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, release=1755695350, vcs-type=git, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.expose-services=, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=minimal rhel9) Nov 27 04:59:26 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:59:26 localhost podman[308334]: 2025-11-27 09:59:26.149021845 +0000 UTC m=+0.238897029 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:59:26 localhost podman[308334]: 2025-11-27 09:59:26.181975362 +0000 UTC m=+0.271850496 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 04:59:26 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:59:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:28 localhost nova_compute[284026]: 2025-11-27 09:59:28.443 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:30 localhost nova_compute[284026]: 2025-11-27 09:59:30.130 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:33 localhost nova_compute[284026]: 2025-11-27 09:59:33.445 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:35 localhost nova_compute[284026]: 2025-11-27 09:59:35.135 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 04:59:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 04:59:35 localhost podman[308393]: 2025-11-27 09:59:35.984432105 +0000 UTC m=+0.084539020 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:59:36 localhost podman[308393]: 2025-11-27 09:59:36.031899304 +0000 UTC m=+0.132006269 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=ovn_controller, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 04:59:36 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 04:59:36 localhost podman[308394]: 2025-11-27 09:59:36.03844496 +0000 UTC m=+0.134528376 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 04:59:36 localhost podman[308394]: 2025-11-27 09:59:36.11788887 +0000 UTC m=+0.213972286 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 04:59:36 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 04:59:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:38 localhost nova_compute[284026]: 2025-11-27 09:59:38.448 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:38 localhost podman[242678]: time="2025-11-27T09:59:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 04:59:38 localhost podman[242678]: @ - - [27/Nov/2025:09:59:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 04:59:38 localhost podman[242678]: @ - - [27/Nov/2025:09:59:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19714 "" "Go-http-client/1.1" Nov 27 04:59:40 localhost nova_compute[284026]: 2025-11-27 09:59:40.172 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e88 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:43 localhost nova_compute[284026]: 2025-11-27 09:59:43.450 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:59:43.568 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 04:59:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:59:43.569 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 04:59:43 localhost ovn_metadata_agent[162087]: 2025-11-27 09:59:43.569 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.136 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.137 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.137 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.137 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.141 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '24190e6d-4ac7-4b53-ad73-6f95170ec577', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.137581', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc86bf30-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '90323e3a359191bf84d65571bce299d06a478e9c2e899b8112880c59c7fb61f0'}]}, 'timestamp': '2025-11-27 09:59:44.141793', '_unique_id': '2d14ffd820ee431aa133c9c69f812bc0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.144 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.144 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5c1966f3-dd7c-45c7-bcbb-b5a86092b394', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.144871', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc875148-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '7a8d3741a4cd5256c73b6472a778918c20f85217041c3663a86041472bdc1ed0'}]}, 'timestamp': '2025-11-27 09:59:44.145626', '_unique_id': '39c01760092845a1bf932bc3cb038aa0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.148 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.178 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.179 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7810dff8-52b8-45bd-9b72-c99d6b3e4508', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.148902', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc8c6d4a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': 'f8b5a6739e3d05bf1dcc40b8d6116c39dd2906166437f958f04334a8c001bea0'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.148902', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc8c84c4-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '7305c49dfbcd56646804d745ec5da12c0cecc7de1557ea3e70118315e01890b7'}]}, 'timestamp': '2025-11-27 09:59:44.179654', '_unique_id': '6928667a032f4661bd3d1d9805c0752b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.181 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.182 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.182 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8200600b-d63a-406c-bc53-ad6aebdbf2ce', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.182068', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc8cfd28-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': 'ba4c3960d96e2e1a55d27c3118aecc6683c1f7d026b00db71f5db223d6a6b4d6'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.182068', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc8d18f8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '64e083999027c93081929fc9a56910ccba52fff2ed45c9e7d31710be77c13452'}]}, 'timestamp': '2025-11-27 09:59:44.183441', '_unique_id': 'e3f926de406c48cb9deabf5f467ef8af'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.186 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.186 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b23f9534-ecac-4fce-9da8-133bd3e8290d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.186341', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc8da2aa-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': 'f5e98d2a3087f3cfba8a4b9f36267282c922f99c9688485aa06bcf06bc4e1823'}]}, 'timestamp': '2025-11-27 09:59:44.186890', '_unique_id': 'b84c55c5b18f48ff9bca919b329e4417'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.187 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.189 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.202 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.203 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c2df98f0-53fc-4874-bf04-80e3d4dbdce5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.189196', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc901b98-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': 'd56368e5374831317a01ef81f66274e3102f56a8752687de36823c37a8dae22f'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.189196', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc903236-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': 'ad8c9c87931c2c1da618d12d8f30cfa17d165458347a53aff19188160a4e2b40'}]}, 'timestamp': '2025-11-27 09:59:44.203767', '_unique_id': 'c23f246097f443f1b75b30e1f3968069'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.204 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.206 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.206 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.206 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c4d29da2-9cd1-4b5e-8147-3144f44220ec', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.206193', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc90a914-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': 'e1147dadfe7f994573bcaaa354796f599fdcff0f57b403efbc94f225f525c2a6'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.206193', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc90bbc0-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': '4b960851113092fb90b85b5949c72bbe1fb6aaf165078f09d56c41ad81ae1eca'}]}, 'timestamp': '2025-11-27 09:59:44.207178', '_unique_id': '712b59df297844f1959f114f2989a46d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.208 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.209 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.209 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.210 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '341f0305-34f2-4423-b8b4-bc158f28a348', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.209510', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc912cb8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '363ee3de71404f3522a5257a644d428c0ba17ef05bfd943987904ff35a505d1f'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.209510', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc913d84-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '98f682cfdf211634d16ad577782208463b4f43504b1001a273ad62415121ec44'}]}, 'timestamp': '2025-11-27 09:59:44.210468', '_unique_id': '481ac52d65194410acc252d877579dc8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.211 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.212 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.212 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 14270000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b27aac8c-cf20-4f81-aa87-1700f6086196', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 14270000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:59:44.213064', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'cc945212-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.418132761, 'message_signature': '44907cc685c4a8ddb835685cea1c557e8be318aef4350316da65e37782c2e378'}]}, 'timestamp': '2025-11-27 09:59:44.230773', '_unique_id': '2b5f3bc21db8403fafccd2dee19e0bf2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '24198c88-b0cd-4b11-9856-2d7675940cac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.233224', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc94c760-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': 'c6f1113aae61a7c5e5f8d96ae9423d9a2662b99ae71b6005322a2785880fa483'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.233224', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc94dc78-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': 'f376bd119f2a716d2aea12fedfe31002f56a752cba516d8e773f12954c20b847'}]}, 'timestamp': '2025-11-27 09:59:44.234206', '_unique_id': '903624ee3b7d4bac83cc1df6dd5a7f95'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ae0bc71d-0f2d-4256-97fb-76212088ffd9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T09:59:44.236610', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'cc954c1c-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.418132761, 'message_signature': '1a8f89b7e9e04ae1b618ce436e03305bb2def16fd46fa48c4b04d217c0ebe768'}]}, 'timestamp': '2025-11-27 09:59:44.237119', '_unique_id': 'b15f127a20fc477d98820b9a5d48c7b2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.239 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '56ed8ffb-b71e-4960-960b-cd115c750727', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.239675', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc95c39a-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': 'f1699eff3631c8b7b09d1f7b34ded3e16fdcfad488f4ae660b994963100faa8a'}]}, 'timestamp': '2025-11-27 09:59:44.240147', '_unique_id': 'c95886d45f2b4d8ca49bf4f9668e4ef3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.242 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '267d569d-2da8-40dd-8fcd-d5377353ce6a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.242614', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc9639a6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': 'e5cdb4015d3f4d9495e5a7ab2a7eb79e4fa236d42401957d9a067b1cdbd76c3a'}]}, 'timestamp': '2025-11-27 09:59:44.243195', '_unique_id': '147796304f864ed288665903b70f35f0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.246 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7eeabda1-ea18-42c1-ac16-0afc6eab4f78', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.245647', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc96adc8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': 'd1b049a7d023117f006b6de35311aa619ebbafc85d9ff9b488496337a41901bf'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.245647', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc96be44-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.377255149, 'message_signature': 'd40497b066540c8fd6eb9bbd984101229f8ec8991dfeb8e0dbba4ef969cfad70'}]}, 'timestamp': '2025-11-27 09:59:44.246567', '_unique_id': '6764bbd740f94aed9d83226ce74bd4cd'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eff50010-e7d2-4fee-9c30-0fb5edb1c8c9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.248896', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc972cc6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '5d2c8af38e8d74f899524a4e28cdf0b0263e72babf0a8a639caf38ace1198cff'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.248896', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc973e8c-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '9ab54ed1a4b0949346116172f428b500ec64265ba667f550b2c5c901c1a93987'}]}, 'timestamp': '2025-11-27 09:59:44.249817', '_unique_id': '997fd7994f3248459273282d71fd0916'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.251 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '54d541ae-ff23-4d5a-9561-13e421fd7ddf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T09:59:44.252129', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'cc97ab74-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '80b8842483f4661dd7e595fc9d83e00c1d80a0b653b83297f983ecc86af824f6'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T09:59:44.252129', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'cc97bd4e-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.336983554, 'message_signature': '86bacaa4ba6c556b720fae2bd1c2d3db6775b11f018e6bf671ab134a02ae74be'}]}, 'timestamp': '2025-11-27 09:59:44.253059', '_unique_id': 'f0810e2b5ae94e3ea7d2bc20ed59e415'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9aeafa0c-08e2-407d-8ee6-67a6479f4178', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.255369', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc982bf8-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '61eb2ec56d599f5904ec6dde1200914fcb8ae9fa9e96783b494776129bc21ae1'}]}, 'timestamp': '2025-11-27 09:59:44.255927', '_unique_id': '9ea7fa8a7b304b3bb9fb4bf5dbee6f8c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.258 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f687a6bb-646a-460d-88a9-1dc50ebc25ce', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.258144', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc9894c6-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '3bc7335b643a871a3eb7882a0a281db09921d1525507aa1390acccda10ab4802'}]}, 'timestamp': '2025-11-27 09:59:44.258637', '_unique_id': 'f11eb109d2234d59aa150251739ffd3a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.261 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8cb6f890-4189-4e2e-bb26-6deaff83d63a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.261159', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc990aaa-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '35b7fb8d8790de4913b671dd57cddbd7d57788a1259387b02831091d99019846'}]}, 'timestamp': '2025-11-27 09:59:44.261663', '_unique_id': '9b698970239848ea86a1d80fd19f60d4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.263 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.263 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9cc0f7be-0e2b-4c32-bae4-8964a461ef85', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.263790', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc996dec-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': '7a89538e196a439dfca063109c4bdd1c1dccaa80218ea896417a539b435790fb'}]}, 'timestamp': '2025-11-27 09:59:44.264080', '_unique_id': '694b8c6aa52e4ac7beeb670e1b0a7031'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.265 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.265 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '226c7be5-83ee-478e-9ba2-4f42a002d89f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T09:59:44.265453', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'cc99aff0-cb77-11f0-8166-fa163e6bfee1', 'monotonic_time': 11814.325632858, 'message_signature': 'a1b8e7943ead5e56408ea2725321121764a63cae83b319a8bb0112c064f94f5f'}]}, 'timestamp': '2025-11-27 09:59:44.265767', '_unique_id': 'e2453f5337954c61936dad082732907c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging yield Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.266 12 ERROR oslo_messaging.notify.messaging Nov 27 04:59:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 09:59:44.267 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 e89: 6 total, 6 up, 6 in Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr handle_mgr_map Activating! Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr handle_mgr_map I am now activating Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537444"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mon metadata", "id": "np0005537444"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537445"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mon metadata", "id": "np0005537445"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon metadata", "id": "np0005537446"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mon metadata", "id": "np0005537446"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mds metadata", "who": "mds.np0005537446.hkzfou"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mds metadata", "who": "mds.np0005537446.hkzfou"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mds metadata", "who": "mds.np0005537444.iqtbiy"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mds metadata", "who": "mds.np0005537444.iqtbiy"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mds metadata", "who": "mds.np0005537445.rkchqo"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mds metadata", "who": "mds.np0005537445.rkchqo"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).mds e17 all = 0 Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).mds e17 all = 0 Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).mds e17 all = 0 Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mgr metadata", "who": "np0005537446.nfulyg", "id": "np0005537446.nfulyg"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mgr metadata", "who": "np0005537446.nfulyg", "id": "np0005537446.nfulyg"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mgr metadata", "who": "np0005537441.xbypnw", "id": "np0005537441.xbypnw"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mgr metadata", "who": "np0005537441.xbypnw", "id": "np0005537441.xbypnw"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mgr metadata", "who": "np0005537444.utkloe", "id": "np0005537444.utkloe"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mgr metadata", "who": "np0005537444.utkloe", "id": "np0005537444.utkloe"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 0} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 0} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 1} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 1} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 2} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 2} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 3} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 3} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 4} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 4} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata", "id": 5} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata", "id": 5} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mds metadata"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mds metadata"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).mds e17 all = 1 Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd metadata"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd metadata"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon metadata"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mon metadata"} : dispatch Nov 27 04:59:44 localhost ceph-mgr[290377]: [balancer DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: balancer Nov 27 04:59:44 localhost ceph-mgr[290377]: [balancer INFO root] Starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_09:59:44 Nov 27 04:59:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 04:59:44 localhost ceph-mgr[290377]: [balancer INFO root] Some PGs (1.000000) are unknown; try again later Nov 27 04:59:44 localhost ceph-mgr[290377]: [cephadm DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost systemd[1]: session-72.scope: Deactivated successfully. Nov 27 04:59:44 localhost systemd[1]: session-72.scope: Consumed 11.535s CPU time. Nov 27 04:59:44 localhost systemd-logind[761]: Session 72 logged out. Waiting for processes to exit. Nov 27 04:59:44 localhost systemd-logind[761]: Removed session 72. Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: cephadm Nov 27 04:59:44 localhost ceph-mgr[290377]: [crash DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: crash Nov 27 04:59:44 localhost ceph-mgr[290377]: [devicehealth DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: devicehealth Nov 27 04:59:44 localhost ceph-mgr[290377]: [devicehealth INFO root] Starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [iostat DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: iostat Nov 27 04:59:44 localhost ceph-mgr[290377]: [nfs DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: nfs Nov 27 04:59:44 localhost ceph-mgr[290377]: [orchestrator DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: orchestrator Nov 27 04:59:44 localhost ceph-mgr[290377]: [pg_autoscaler DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: pg_autoscaler Nov 27 04:59:44 localhost ceph-mgr[290377]: [progress DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: progress Nov 27 04:59:44 localhost ceph-mgr[290377]: [progress INFO root] Loading... Nov 27 04:59:44 localhost ceph-mgr[290377]: [progress INFO root] Loaded [, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ] historic events Nov 27 04:59:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 04:59:44 localhost ceph-mgr[290377]: [progress INFO root] Loaded OSDMap, ready. Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] recovery thread starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] starting setup Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: rbd_support Nov 27 04:59:44 localhost ceph-mgr[290377]: [restful DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: restful Nov 27 04:59:44 localhost ceph-mgr[290377]: [restful INFO root] server_addr: :: server_port: 8003 Nov 27 04:59:44 localhost ceph-mgr[290377]: [restful WARNING root] server not running: no certificate configured Nov 27 04:59:44 localhost ceph-mgr[290377]: [status DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: status Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/mirror_snapshot_schedule"} v 0) Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/mirror_snapshot_schedule"} : dispatch Nov 27 04:59:44 localhost ceph-mgr[290377]: [telemetry DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: telemetry Nov 27 04:59:44 localhost ceph-mgr[290377]: [volumes DEBUG root] setting log level based on debug_mgr: INFO (2/5) Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] PerfHandler: starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: vms, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 04:59:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 04:59:44 localhost ceph-mgr[290377]: mgr load Constructed class from module: volumes Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: volumes, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: images, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_task_task: backups, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.540+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.540+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.540+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.540+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.540+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/trash_purge_schedule"} v 0) Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TaskHandler: starting Nov 27 04:59:44 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/trash_purge_schedule"} : dispatch Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.542+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.542+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.542+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.542+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T09:59:44.542+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: starting Nov 27 04:59:44 localhost ceph-mgr[290377]: [rbd_support INFO root] setup complete Nov 27 04:59:44 localhost ceph-mon[305520]: from='client.? 172.18.0.200:0/3900846710' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: from='client.? ' entity='client.admin' cmd={"prefix": "mgr fail"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: Activating manager daemon np0005537446.nfulyg Nov 27 04:59:44 localhost ceph-mon[305520]: from='client.? ' entity='client.admin' cmd='[{"prefix": "mgr fail"}]': finished Nov 27 04:59:44 localhost ceph-mon[305520]: Manager daemon np0005537446.nfulyg is now available Nov 27 04:59:44 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/mirror_snapshot_schedule"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/mirror_snapshot_schedule"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/trash_purge_schedule"} : dispatch Nov 27 04:59:44 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/np0005537446.nfulyg/trash_purge_schedule"} : dispatch Nov 27 04:59:44 localhost sshd[308579]: main: sshd: ssh-rsa algorithm is disabled Nov 27 04:59:44 localhost systemd-logind[761]: New session 73 of user ceph-admin. Nov 27 04:59:44 localhost systemd[1]: Started Session 73 of User ceph-admin. Nov 27 04:59:45 localhost nova_compute[284026]: 2025-11-27 09:59:45.172 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v3: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 04:59:45 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:59:45] ENGINE Bus STARTING Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:59:45] ENGINE Bus STARTING Nov 27 04:59:45 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:59:45] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:59:45] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:59:45 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:59:45] ENGINE Client ('172.18.0.108', 43610) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:59:45] ENGINE Client ('172.18.0.108', 43610) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:59:45 localhost podman[308695]: 2025-11-27 09:59:45.87977275 +0000 UTC m=+0.097787006 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, build-date=2025-09-24T08:57:55, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.openshift.expose-services=, release=553, GIT_BRANCH=main, maintainer=Guillaume Abrioux , GIT_CLEAN=True, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, com.redhat.component=rhceph-container, ceph=True, description=Red Hat Ceph Storage 7, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.license_terms=https://www.redhat.com/agreements, version=7, RELEASE=main, vcs-type=git, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.tags=rhceph ceph, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64) Nov 27 04:59:45 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:59:45] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:59:45] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:59:45 localhost ceph-mgr[290377]: [cephadm INFO cherrypy.error] [27/Nov/2025:09:59:45] ENGINE Bus STARTED Nov 27 04:59:45 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : [27/Nov/2025:09:59:45] ENGINE Bus STARTED Nov 27 04:59:45 localhost podman[308695]: 2025-11-27 09:59:45.984967545 +0000 UTC m=+0.202981841 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, maintainer=Guillaume Abrioux , version=7, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, distribution-scope=public, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_CLEAN=True, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_BRANCH=main, io.openshift.tags=rhceph ceph, io.buildah.version=1.33.12, io.openshift.expose-services=, release=553, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, ceph=True, description=Red Hat Ceph Storage 7, com.redhat.component=rhceph-container, vcs-type=git) Nov 27 04:59:46 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v4: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:59:46 localhost ceph-mgr[290377]: [devicehealth INFO root] Check health Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:59:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:47 localhost ceph-mon[305520]: [27/Nov/2025:09:59:45] ENGINE Bus STARTING Nov 27 04:59:47 localhost ceph-mon[305520]: [27/Nov/2025:09:59:45] ENGINE Serving on https://172.18.0.108:7150 Nov 27 04:59:47 localhost ceph-mon[305520]: [27/Nov/2025:09:59:45] ENGINE Client ('172.18.0.108', 43610) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') Nov 27 04:59:47 localhost ceph-mon[305520]: [27/Nov/2025:09:59:45] ENGINE Serving on http://172.18.0.108:8765 Nov 27 04:59:47 localhost ceph-mon[305520]: [27/Nov/2025:09:59:45] ENGINE Bus STARTED Nov 27 04:59:47 localhost ceph-mon[305520]: Health check cleared: CEPHADM_STRAY_DAEMON (was: 2 stray daemon(s) not managed by cephadm) Nov 27 04:59:47 localhost ceph-mon[305520]: Health check cleared: CEPHADM_STRAY_HOST (was: 2 stray host(s) with 2 daemon(s) not managed by cephadm) Nov 27 04:59:47 localhost ceph-mon[305520]: Cluster is now healthy Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:59:47 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:59:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} v 0) Nov 27 04:59:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:59:47 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:47 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:59:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:59:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 04:59:47 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:47 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 04:59:48 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 04:59:48 localhost podman[308985]: 2025-11-27 09:59:48.353623072 +0000 UTC m=+0.103807898 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v5: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 04:59:48 localhost podman[308985]: 2025-11-27 09:59:48.385329696 +0000 UTC m=+0.135514512 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 04:59:48 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 04:59:48 localhost podman[308986]: 2025-11-27 09:59:48.410219917 +0000 UTC m=+0.160579238 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, config_id=multipathd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, org.label-schema.license=GPLv2) Nov 27 04:59:48 localhost podman[308986]: 2025-11-27 09:59:48.446571887 +0000 UTC m=+0.196931188 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=multipathd) Nov 27 04:59:48 localhost nova_compute[284026]: 2025-11-27 09:59:48.454 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:48 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2527209295' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 04:59:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2527209295' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 04:59:48 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:48 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/etc/ceph/ceph.conf Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 04:59:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 04:59:49 localhost ceph-mgr[290377]: mgr.server handle_open ignoring open from mgr.np0005537445.wmiblp 172.18.0.107:0/702277995; not ready for session (expect reconnect) Nov 27 04:59:49 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:49 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:49 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:49 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:49 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:49 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.conf Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/etc/ceph/ceph.client.admin.keyring Nov 27 04:59:50 localhost nova_compute[284026]: 2025-11-27 09:59:50.175 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mgr metadata", "who": "np0005537445.wmiblp", "id": "np0005537445.wmiblp"} v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "mgr metadata", "who": "np0005537445.wmiblp", "id": "np0005537445.wmiblp"} : dispatch Nov 27 04:59:50 localhost ceph-mgr[290377]: [cephadm INFO cephadm.serve] Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:50 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v6: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 04:59:50 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v7: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 37 KiB/s rd, 0 B/s wr, 20 op/s Nov 27 04:59:50 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 80e67a57-4aba-4c73-bac1-eead1d99d4f1 (Updating node-proxy deployment (+3 -> 3)) Nov 27 04:59:50 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 80e67a57-4aba-4c73-bac1-eead1d99d4f1 (Updating node-proxy deployment (+3 -> 3)) Nov 27 04:59:50 localhost ceph-mgr[290377]: [progress INFO root] Completed event 80e67a57-4aba-4c73-bac1-eead1d99d4f1 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 04:59:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 04:59:50 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 04:59:51 localhost ceph-mon[305520]: Updating np0005537445.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:51 localhost ceph-mon[305520]: Updating np0005537444.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:51 localhost ceph-mon[305520]: Updating np0005537446.localdomain:/var/lib/ceph/e83f3b0c-4090-52df-95d4-ad9be8516692/config/ceph.client.admin.keyring Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 04:59:51 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 04:59:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 04:59:51 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:59:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 04:59:51 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev a61e0d14-2f11-4aff-b3a0-8e66a6dd8b9b (Updating node-proxy deployment (+3 -> 3)) Nov 27 04:59:51 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev a61e0d14-2f11-4aff-b3a0-8e66a6dd8b9b (Updating node-proxy deployment (+3 -> 3)) Nov 27 04:59:51 localhost ceph-mgr[290377]: [progress INFO root] Completed event a61e0d14-2f11-4aff-b3a0-8e66a6dd8b9b (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 04:59:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 04:59:51 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 04:59:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:52 localhost ceph-mon[305520]: Health check failed: 1 stray daemon(s) not managed by cephadm (CEPHADM_STRAY_DAEMON) Nov 27 04:59:52 localhost ceph-mon[305520]: Health check failed: 1 stray host(s) with 1 daemon(s) not managed by cephadm (CEPHADM_STRAY_HOST) Nov 27 04:59:52 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 04:59:52 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:52 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v8: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 27 KiB/s rd, 0 B/s wr, 15 op/s Nov 27 04:59:53 localhost nova_compute[284026]: 2025-11-27 09:59:53.457 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:54 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 04:59:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 04:59:54 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v9: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 0 B/s wr, 12 op/s Nov 27 04:59:55 localhost nova_compute[284026]: 2025-11-27 09:59:55.207 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:55 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #16. Immutable memtables: 0. Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.527718) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 5] Flushing memtable with next log file: 16 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595527792, "job": 5, "event": "flush_started", "num_memtables": 1, "num_entries": 2514, "num_deletes": 256, "total_data_size": 9947357, "memory_usage": 10381216, "flush_reason": "Manual Compaction"} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 5] Level-0 flush table #17: started Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595563976, "cf_name": "default", "job": 5, "event": "table_file_creation", "file_number": 17, "file_size": 6297147, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 12159, "largest_seqno": 14672, "table_properties": {"data_size": 6286749, "index_size": 6455, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2885, "raw_key_size": 25506, "raw_average_key_size": 22, "raw_value_size": 6264529, "raw_average_value_size": 5475, "num_data_blocks": 278, "num_entries": 1144, "num_filter_entries": 1144, "num_deletions": 255, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237509, "oldest_key_time": 1764237509, "file_creation_time": 1764237595, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 17, "seqno_to_time_mapping": "N/A"}} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 5] Flush lasted 36314 microseconds, and 12434 cpu microseconds. Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.564041) [db/flush_job.cc:967] [default] [JOB 5] Level-0 flush table #17: 6297147 bytes OK Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.564067) [db/memtable_list.cc:519] [default] Level-0 commit table #17 started Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.583582) [db/memtable_list.cc:722] [default] Level-0 commit table #17: memtable #1 done Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.583608) EVENT_LOG_v1 {"time_micros": 1764237595583602, "job": 5, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.583632) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 5] Try to delete WAL files size 9935299, prev total WAL file size 9935299, number of live WAL files 2. Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000013.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.586308) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003131323935' seq:72057594037927935, type:22 .. '7061786F73003131353437' seq:0, type:0; will stop at (end) Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 6] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 5 Base level 0, inputs: [17(6149KB)], [15(17MB)] Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595586364, "job": 6, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [17], "files_L6": [15], "score": -1, "input_data_size": 24461107, "oldest_snapshot_seqno": -1} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 6] Generated table #18: 12162 keys, 21660917 bytes, temperature: kUnknown Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595760055, "cf_name": "default", "job": 6, "event": "table_file_creation", "file_number": 18, "file_size": 21660917, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 21588884, "index_size": 40508, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 30469, "raw_key_size": 324905, "raw_average_key_size": 26, "raw_value_size": 21379216, "raw_average_value_size": 1757, "num_data_blocks": 1553, "num_entries": 12162, "num_filter_entries": 12162, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237595, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.760428) [db/compaction/compaction_job.cc:1663] [default] [JOB 6] Compacted 1@0 + 1@6 files to L6 => 21660917 bytes Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.762205) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 140.8 rd, 124.7 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(6.0, 17.3 +0.0 blob) out(20.7 +0.0 blob), read-write-amplify(7.3) write-amplify(3.4) OK, records in: 12710, records dropped: 548 output_compression: NoCompression Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.762246) EVENT_LOG_v1 {"time_micros": 1764237595762228, "job": 6, "event": "compaction_finished", "compaction_time_micros": 173766, "compaction_time_cpu_micros": 55131, "output_level": 6, "num_output_files": 1, "total_output_size": 21660917, "num_input_records": 12710, "num_output_records": 12162, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000017.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595763325, "job": 6, "event": "table_file_deletion", "file_number": 17} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000015.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237595766261, "job": 6, "event": "table_file_deletion", "file_number": 15} Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.586224) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.766393) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.766402) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.766405) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.766408) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-09:59:55.766411) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 04:59:55 localhost openstack_network_exporter[244641]: ERROR 09:59:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 04:59:55 localhost openstack_network_exporter[244641]: ERROR 09:59:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:59:55 localhost openstack_network_exporter[244641]: ERROR 09:59:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 04:59:55 localhost openstack_network_exporter[244641]: ERROR 09:59:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 04:59:55 localhost openstack_network_exporter[244641]: Nov 27 04:59:55 localhost openstack_network_exporter[244641]: ERROR 09:59:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 04:59:55 localhost openstack_network_exporter[244641]: Nov 27 04:59:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 04:59:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 04:59:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 04:59:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 04:59:56 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v10: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 04:59:56 localhost podman[309680]: 2025-11-27 09:59:56.985897151 +0000 UTC m=+0.077754865 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_id=edpm, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 04:59:56 localhost systemd[1]: tmp-crun.cmI64b.mount: Deactivated successfully. Nov 27 04:59:57 localhost podman[309682]: 2025-11-27 09:59:57.002710165 +0000 UTC m=+0.085988418 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, build-date=2025-08-20T13:12:41, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc., distribution-scope=public, release=1755695350, container_name=openstack_network_exporter, vendor=Red Hat, Inc., name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, com.redhat.component=ubi9-minimal-container, architecture=x86_64, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 04:59:57 localhost podman[309681]: 2025-11-27 09:59:57.050119702 +0000 UTC m=+0.137002953 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 04:59:57 localhost podman[309680]: 2025-11-27 09:59:57.055201079 +0000 UTC m=+0.147058813 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 04:59:57 localhost podman[309682]: 2025-11-27 09:59:57.07007025 +0000 UTC m=+0.153348473 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.component=ubi9-minimal-container, distribution-scope=public, maintainer=Red Hat, Inc., config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, container_name=openstack_network_exporter, name=ubi9-minimal, build-date=2025-08-20T13:12:41, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, io.openshift.expose-services=, managed_by=edpm_ansible, vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 04:59:57 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 04:59:57 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 04:59:57 localhost podman[309681]: 2025-11-27 09:59:57.090032198 +0000 UTC m=+0.176915429 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 04:59:57 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 04:59:58 localhost nova_compute[284026]: 2025-11-27 09:59:58.494 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 04:59:58 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v11: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 05:00:00 localhost nova_compute[284026]: 2025-11-27 10:00:00.210 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:00 localhost ceph-mon[305520]: Health detail: HEALTH_WARN 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 05:00:00 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm Nov 27 05:00:00 localhost ceph-mon[305520]: stray daemon mgr.np0005537441.xbypnw on host np0005537441.localdomain not managed by cephadm Nov 27 05:00:00 localhost ceph-mon[305520]: [WRN] CEPHADM_STRAY_HOST: 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 05:00:00 localhost ceph-mon[305520]: stray host np0005537441.localdomain has 1 stray daemons: ['mgr.np0005537441.xbypnw'] Nov 27 05:00:00 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v12: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 0 B/s wr, 10 op/s Nov 27 05:00:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:02 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v13: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:03 localhost nova_compute[284026]: 2025-11-27 10:00:03.497 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:04 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v14: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:05 localhost nova_compute[284026]: 2025-11-27 10:00:05.253 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:00:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:00:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:06 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v15: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:06 localhost podman[309739]: 2025-11-27 10:00:06.998116177 +0000 UTC m=+0.089506843 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, managed_by=edpm_ansible, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller) Nov 27 05:00:07 localhost podman[309740]: 2025-11-27 10:00:07.045478413 +0000 UTC m=+0.132795929 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:00:07 localhost podman[309739]: 2025-11-27 10:00:07.061962738 +0000 UTC m=+0.153353354 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0) Nov 27 05:00:07 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:00:07 localhost podman[309740]: 2025-11-27 10:00:07.078996747 +0000 UTC m=+0.166314253 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:00:07 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:00:08 localhost nova_compute[284026]: 2025-11-27 10:00:08.540 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:08 localhost podman[242678]: time="2025-11-27T10:00:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:00:08 localhost podman[242678]: @ - - [27/Nov/2025:10:00:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:00:08 localhost podman[242678]: @ - - [27/Nov/2025:10:00:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19711 "" "Go-http-client/1.1" Nov 27 05:00:08 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v16: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:10 localhost nova_compute[284026]: 2025-11-27 10:00:10.252 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:10 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v17: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:12 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v18: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:13 localhost nova_compute[284026]: 2025-11-27 10:00:13.543 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:14 localhost nova_compute[284026]: 2025-11-27 10:00:14.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:14 localhost nova_compute[284026]: 2025-11-27 10:00:14.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:14 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v19: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:15 localhost nova_compute[284026]: 2025-11-27 10:00:15.256 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:16 localhost nova_compute[284026]: 2025-11-27 10:00:16.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:16 localhost nova_compute[284026]: 2025-11-27 10:00:16.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:16 localhost nova_compute[284026]: 2025-11-27 10:00:16.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:16 localhost nova_compute[284026]: 2025-11-27 10:00:16.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:00:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:16 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v20: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:18 localhost nova_compute[284026]: 2025-11-27 10:00:18.581 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:00:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:00:18 localhost podman[309784]: 2025-11-27 10:00:18.983135373 +0000 UTC m=+0.079638257 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:00:18 localhost podman[309784]: 2025-11-27 10:00:18.988438216 +0000 UTC m=+0.084941100 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:00:18 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v21: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:19 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:00:19 localhost podman[309785]: 2025-11-27 10:00:19.046632463 +0000 UTC m=+0.140616250 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:00:19 localhost podman[309785]: 2025-11-27 10:00:19.06206901 +0000 UTC m=+0.156052757 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:00:19 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.846 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.846 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.847 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.847 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:00:19 localhost nova_compute[284026]: 2025-11-27 10:00:19.848 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.259 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:00:20 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1783526779' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.355 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.508s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.691 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.691 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.886 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.887 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11411MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.888 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:00:20 localhost nova_compute[284026]: 2025-11-27 10:00:20.888 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:00:20 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v22: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.118 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.119 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.119 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.154 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:00:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:00:21 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1893878099' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.604 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.449s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.611 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.664 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.667 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:00:21 localhost nova_compute[284026]: 2025-11-27 10:00:21.667 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.780s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:00:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.669 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.943 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.943 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.944 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:00:22 localhost nova_compute[284026]: 2025-11-27 10:00:22.945 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:00:22 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v23: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:23 localhost nova_compute[284026]: 2025-11-27 10:00:23.584 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:24 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v24: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:25 localhost nova_compute[284026]: 2025-11-27 10:00:25.262 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:25 localhost nova_compute[284026]: 2025-11-27 10:00:25.596 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:00:25 localhost openstack_network_exporter[244641]: ERROR 10:00:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:00:25 localhost openstack_network_exporter[244641]: ERROR 10:00:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:00:25 localhost openstack_network_exporter[244641]: ERROR 10:00:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:00:25 localhost openstack_network_exporter[244641]: ERROR 10:00:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:00:25 localhost openstack_network_exporter[244641]: Nov 27 05:00:25 localhost openstack_network_exporter[244641]: ERROR 10:00:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:00:25 localhost openstack_network_exporter[244641]: Nov 27 05:00:26 localhost nova_compute[284026]: 2025-11-27 10:00:26.433 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:00:26 localhost nova_compute[284026]: 2025-11-27 10:00:26.433 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:00:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:26 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v25: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:00:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:00:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:00:27 localhost podman[309866]: 2025-11-27 10:00:27.995255199 +0000 UTC m=+0.090605234 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:00:28 localhost podman[309866]: 2025-11-27 10:00:28.008070373 +0000 UTC m=+0.103420408 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:00:28 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:00:28 localhost podman[309867]: 2025-11-27 10:00:28.095911351 +0000 UTC m=+0.188122261 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:00:28 localhost podman[309867]: 2025-11-27 10:00:28.10890461 +0000 UTC m=+0.201115520 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:00:28 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:00:28 localhost podman[309868]: 2025-11-27 10:00:28.203727556 +0000 UTC m=+0.293012076 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, build-date=2025-08-20T13:12:41, architecture=x86_64, vcs-type=git, version=9.6, name=ubi9-minimal, vendor=Red Hat, Inc., io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal) Nov 27 05:00:28 localhost podman[309868]: 2025-11-27 10:00:28.222263046 +0000 UTC m=+0.311547586 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, build-date=2025-08-20T13:12:41, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, container_name=openstack_network_exporter, distribution-scope=public, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, vendor=Red Hat, Inc., io.buildah.version=1.33.7, io.openshift.expose-services=) Nov 27 05:00:28 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:00:28 localhost nova_compute[284026]: 2025-11-27 10:00:28.607 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:28 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v26: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:30 localhost nova_compute[284026]: 2025-11-27 10:00:30.264 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:30 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v27: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:32 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v28: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:33 localhost nova_compute[284026]: 2025-11-27 10:00:33.611 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:34 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v29: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:35 localhost nova_compute[284026]: 2025-11-27 10:00:35.269 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:36 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v30: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:00:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:00:37 localhost podman[309927]: 2025-11-27 10:00:37.983920668 +0000 UTC m=+0.082916435 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:00:38 localhost podman[309928]: 2025-11-27 10:00:38.055255281 +0000 UTC m=+0.148242186 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:00:38 localhost podman[309927]: 2025-11-27 10:00:38.063420191 +0000 UTC m=+0.162416018 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:00:38 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:00:38 localhost podman[309928]: 2025-11-27 10:00:38.090909331 +0000 UTC m=+0.183896216 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:00:38 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:00:38 localhost nova_compute[284026]: 2025-11-27 10:00:38.615 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:38 localhost podman[242678]: time="2025-11-27T10:00:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:00:38 localhost podman[242678]: @ - - [27/Nov/2025:10:00:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:00:38 localhost podman[242678]: @ - - [27/Nov/2025:10:00:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19715 "" "Go-http-client/1.1" Nov 27 05:00:38 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v31: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:40 localhost nova_compute[284026]: 2025-11-27 10:00:40.270 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:40 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v32: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:42 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v33: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:00:43.569 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:00:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:00:43.570 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:00:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:00:43.571 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:00:43 localhost nova_compute[284026]: 2025-11-27 10:00:43.617 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:00:44 Nov 27 05:00:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:00:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:00:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['.mgr', 'backups', 'manila_metadata', 'vms', 'images', 'volumes', 'manila_data'] Nov 27 05:00:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.0014449417225013959 of space, bias 1.0, pg target 0.2885066972594454 quantized to 32 (current 32) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:00:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.0021774090359203426 quantized to 16 (current 16) Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:00:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:00:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:00:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v34: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:45 localhost nova_compute[284026]: 2025-11-27 10:00:45.277 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #19. Immutable memtables: 0. Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.339021) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 7] Flushing memtable with next log file: 19 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646339101, "job": 7, "event": "flush_started", "num_memtables": 1, "num_entries": 786, "num_deletes": 250, "total_data_size": 767043, "memory_usage": 780512, "flush_reason": "Manual Compaction"} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 7] Level-0 flush table #20: started Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646345036, "cf_name": "default", "job": 7, "event": "table_file_creation", "file_number": 20, "file_size": 331273, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 14677, "largest_seqno": 15458, "table_properties": {"data_size": 328336, "index_size": 856, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1029, "raw_key_size": 8150, "raw_average_key_size": 20, "raw_value_size": 322094, "raw_average_value_size": 815, "num_data_blocks": 38, "num_entries": 395, "num_filter_entries": 395, "num_deletions": 250, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237596, "oldest_key_time": 1764237596, "file_creation_time": 1764237646, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 20, "seqno_to_time_mapping": "N/A"}} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 7] Flush lasted 6045 microseconds, and 2387 cpu microseconds. Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.345080) [db/flush_job.cc:967] [default] [JOB 7] Level-0 flush table #20: 331273 bytes OK Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.345101) [db/memtable_list.cc:519] [default] Level-0 commit table #20 started Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.350648) [db/memtable_list.cc:722] [default] Level-0 commit table #20: memtable #1 done Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.350671) EVENT_LOG_v1 {"time_micros": 1764237646350665, "job": 7, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.350692) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 7] Try to delete WAL files size 762899, prev total WAL file size 763223, number of live WAL files 2. Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000016.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.352334) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6D6772737461740033373535' seq:72057594037927935, type:22 .. '6D6772737461740034303036' seq:0, type:0; will stop at (end) Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 8] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 7 Base level 0, inputs: [20(323KB)], [18(20MB)] Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646352456, "job": 8, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [20], "files_L6": [18], "score": -1, "input_data_size": 21992190, "oldest_snapshot_seqno": -1} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 8] Generated table #21: 12062 keys, 19970429 bytes, temperature: kUnknown Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646501249, "cf_name": "default", "job": 8, "event": "table_file_creation", "file_number": 21, "file_size": 19970429, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19903339, "index_size": 35821, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 30213, "raw_key_size": 323079, "raw_average_key_size": 26, "raw_value_size": 19699624, "raw_average_value_size": 1633, "num_data_blocks": 1358, "num_entries": 12062, "num_filter_entries": 12062, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237646, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 21, "seqno_to_time_mapping": "N/A"}} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.501667) [db/compaction/compaction_job.cc:1663] [default] [JOB 8] Compacted 1@0 + 1@6 files to L6 => 19970429 bytes Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.503603) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 147.7 rd, 134.1 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.3, 20.7 +0.0 blob) out(19.0 +0.0 blob), read-write-amplify(126.7) write-amplify(60.3) OK, records in: 12557, records dropped: 495 output_compression: NoCompression Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.503647) EVENT_LOG_v1 {"time_micros": 1764237646503629, "job": 8, "event": "compaction_finished", "compaction_time_micros": 148909, "compaction_time_cpu_micros": 52939, "output_level": 6, "num_output_files": 1, "total_output_size": 19970429, "num_input_records": 12557, "num_output_records": 12062, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000020.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646503889, "job": 8, "event": "table_file_deletion", "file_number": 20} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000018.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237646507001, "job": 8, "event": "table_file_deletion", "file_number": 18} Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.352113) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.507158) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.507167) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.507170) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.507173) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:00:46.507176) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:00:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v35: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:48 localhost nova_compute[284026]: 2025-11-27 10:00:48.659 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v36: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:00:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:00:49 localhost podman[309976]: 2025-11-27 10:00:49.989563679 +0000 UTC m=+0.078022643 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:00:49 localhost podman[309976]: 2025-11-27 10:00:49.997897604 +0000 UTC m=+0.086356598 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.schema-version=1.0, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:00:50 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:00:50 localhost podman[309975]: 2025-11-27 10:00:50.046855692 +0000 UTC m=+0.136283452 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent) Nov 27 05:00:50 localhost podman[309975]: 2025-11-27 10:00:50.052239378 +0000 UTC m=+0.141667118 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:00:50 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:00:50 localhost nova_compute[284026]: 2025-11-27 10:00:50.278 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v37: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:00:52 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:00:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:00:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:00:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:00:52 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 17a62067-ccfd-4786-a0a8-21b51fe8e48a (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:00:52 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 17a62067-ccfd-4786-a0a8-21b51fe8e48a (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:00:52 localhost ceph-mgr[290377]: [progress INFO root] Completed event 17a62067-ccfd-4786-a0a8-21b51fe8e48a (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:00:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:00:52 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:00:52 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:00:52 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:00:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v38: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:53 localhost nova_compute[284026]: 2025-11-27 10:00:53.660 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:54 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:00:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:00:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v39: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:55 localhost nova_compute[284026]: 2025-11-27 10:00:55.282 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:55 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:00:55 localhost openstack_network_exporter[244641]: ERROR 10:00:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:00:55 localhost openstack_network_exporter[244641]: ERROR 10:00:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:00:55 localhost openstack_network_exporter[244641]: ERROR 10:00:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:00:55 localhost openstack_network_exporter[244641]: ERROR 10:00:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:00:55 localhost openstack_network_exporter[244641]: Nov 27 05:00:55 localhost openstack_network_exporter[244641]: ERROR 10:00:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:00:55 localhost openstack_network_exporter[244641]: Nov 27 05:00:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:00:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v40: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:58 localhost nova_compute[284026]: 2025-11-27 10:00:58.688 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:00:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:00:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:00:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:00:58 localhost podman[310100]: 2025-11-27 10:00:58.997938235 +0000 UTC m=+0.088310591 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:00:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v41: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:00:59 localhost podman[310100]: 2025-11-27 10:00:59.032865766 +0000 UTC m=+0.123238082 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:00:59 localhost systemd[1]: tmp-crun.4a4Syk.mount: Deactivated successfully. Nov 27 05:00:59 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:00:59 localhost podman[310101]: 2025-11-27 10:00:59.055991458 +0000 UTC m=+0.143283101 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, version=9.6, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, distribution-scope=public, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, io.openshift.expose-services=, vcs-type=git, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., managed_by=edpm_ansible, release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, io.openshift.tags=minimal rhel9) Nov 27 05:00:59 localhost podman[310101]: 2025-11-27 10:00:59.098985668 +0000 UTC m=+0.186277291 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, managed_by=edpm_ansible, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vcs-type=git, container_name=openstack_network_exporter, architecture=x86_64, distribution-scope=public, url=https://catalog.redhat.com/en/search?searchType=containers, io.buildah.version=1.33.7, version=9.6, maintainer=Red Hat, Inc., build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc.) Nov 27 05:00:59 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:00:59 localhost podman[310099]: 2025-11-27 10:00:59.147880525 +0000 UTC m=+0.241265262 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, container_name=ceilometer_agent_compute) Nov 27 05:00:59 localhost podman[310099]: 2025-11-27 10:00:59.16293076 +0000 UTC m=+0.256315537 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.vendor=CentOS) Nov 27 05:00:59 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:01:00 localhost nova_compute[284026]: 2025-11-27 10:01:00.287 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v42: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v43: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:03 localhost nova_compute[284026]: 2025-11-27 10:01:03.691 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v44: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:05 localhost nova_compute[284026]: 2025-11-27 10:01:05.291 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v45: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:08 localhost nova_compute[284026]: 2025-11-27 10:01:08.694 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:08 localhost podman[242678]: time="2025-11-27T10:01:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:01:08 localhost podman[242678]: @ - - [27/Nov/2025:10:01:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:01:08 localhost podman[242678]: @ - - [27/Nov/2025:10:01:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19708 "" "Go-http-client/1.1" Nov 27 05:01:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:01:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:01:09 localhost podman[310171]: 2025-11-27 10:01:08.999540902 +0000 UTC m=+0.090256783 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:01:09 localhost podman[310171]: 2025-11-27 10:01:09.009441889 +0000 UTC m=+0.100157790 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:01:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v46: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:09 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:01:09 localhost systemd[1]: tmp-crun.WW7CnA.mount: Deactivated successfully. Nov 27 05:01:09 localhost podman[310170]: 2025-11-27 10:01:09.106860254 +0000 UTC m=+0.196423794 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, container_name=ovn_controller, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:01:09 localhost podman[310170]: 2025-11-27 10:01:09.173091799 +0000 UTC m=+0.262655329 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:01:09 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:01:10 localhost nova_compute[284026]: 2025-11-27 10:01:10.303 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v47: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v48: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:13 localhost nova_compute[284026]: 2025-11-27 10:01:13.696 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', )] Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', )] Nov 27 05:01:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:01:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v49: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:15 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:15.277 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=7, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=6) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:01:15 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:15.279 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:01:15 localhost nova_compute[284026]: 2025-11-27 10:01:15.278 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:15 localhost nova_compute[284026]: 2025-11-27 10:01:15.305 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:15 localhost nova_compute[284026]: 2025-11-27 10:01:15.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:15 localhost nova_compute[284026]: 2025-11-27 10:01:15.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:16 localhost nova_compute[284026]: 2025-11-27 10:01:16.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:16 localhost nova_compute[284026]: 2025-11-27 10:01:16.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e89 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v50: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 255 B/s wr, 0 op/s Nov 27 05:01:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:18.281 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '7'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:01:18 localhost nova_compute[284026]: 2025-11-27 10:01:18.721 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:18 localhost nova_compute[284026]: 2025-11-27 10:01:18.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:18 localhost nova_compute[284026]: 2025-11-27 10:01:18.731 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:01:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v51: 177 pgs: 177 active+clean; 105 MiB data, 584 MiB used, 41 GiB / 42 GiB avail; 255 B/s wr, 0 op/s Nov 27 05:01:19 localhost nova_compute[284026]: 2025-11-27 10:01:19.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:20 localhost nova_compute[284026]: 2025-11-27 10:01:20.308 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:20 localhost nova_compute[284026]: 2025-11-27 10:01:20.726 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v52: 177 pgs: 177 active+clean; 125 MiB data, 663 MiB used, 41 GiB / 42 GiB avail; 8.2 KiB/s rd, 1.7 MiB/s wr, 12 op/s Nov 27 05:01:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:01:21 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:01:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e90 e90: 6 total, 6 up, 6 in Nov 27 05:01:21 localhost systemd[1]: tmp-crun.U0Nmi9.mount: Deactivated successfully. Nov 27 05:01:21 localhost podman[310220]: 2025-11-27 10:01:21.186777415 +0000 UTC m=+0.097446657 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, org.label-schema.build-date=20251125, tcib_managed=true, config_id=multipathd) Nov 27 05:01:21 localhost podman[310220]: 2025-11-27 10:01:21.200082844 +0000 UTC m=+0.110752076 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:01:21 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:01:21 localhost podman[310219]: 2025-11-27 10:01:21.156217662 +0000 UTC m=+0.074282292 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, io.buildah.version=1.41.3) Nov 27 05:01:21 localhost podman[310219]: 2025-11-27 10:01:21.290004387 +0000 UTC m=+0.208068937 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:01:21 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.754 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:01:21 localhost nova_compute[284026]: 2025-11-27 10:01:21.755 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:01:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e90 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:01:22 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/953033181' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.293 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.538s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.365 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.366 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.618 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.620 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11409MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.620 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.620 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.877 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.878 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.878 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:01:22 localhost nova_compute[284026]: 2025-11-27 10:01:22.910 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:01:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v54: 177 pgs: 177 active+clean; 125 MiB data, 663 MiB used, 41 GiB / 42 GiB avail; 9.9 KiB/s rd, 2.0 MiB/s wr, 14 op/s Nov 27 05:01:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 e91: 6 total, 6 up, 6 in Nov 27 05:01:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:01:23 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2304118015' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.398 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.489s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.405 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.426 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.428 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.428 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.808s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:01:23 localhost nova_compute[284026]: 2025-11-27 10:01:23.725 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v56: 177 pgs: 177 active+clean; 125 MiB data, 663 MiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 2.6 MiB/s wr, 18 op/s Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.311 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.430 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.431 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.431 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:01:25 localhost openstack_network_exporter[244641]: ERROR 10:01:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:01:25 localhost openstack_network_exporter[244641]: ERROR 10:01:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:01:25 localhost openstack_network_exporter[244641]: Nov 27 05:01:25 localhost openstack_network_exporter[244641]: ERROR 10:01:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:01:25 localhost openstack_network_exporter[244641]: Nov 27 05:01:25 localhost openstack_network_exporter[244641]: ERROR 10:01:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:01:25 localhost openstack_network_exporter[244641]: ERROR 10:01:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.993 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.994 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.994 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:01:25 localhost nova_compute[284026]: 2025-11-27 10:01:25.995 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:01:26 localhost nova_compute[284026]: 2025-11-27 10:01:26.594 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:01:26 localhost nova_compute[284026]: 2025-11-27 10:01:26.809 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:01:26 localhost nova_compute[284026]: 2025-11-27 10:01:26.810 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:01:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v57: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 5.1 MiB/s wr, 47 op/s Nov 27 05:01:28 localhost nova_compute[284026]: 2025-11-27 10:01:28.769 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v58: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 2.6 MiB/s wr, 29 op/s Nov 27 05:01:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:01:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:01:29 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:01:29 localhost systemd[1]: tmp-crun.X4BQk0.mount: Deactivated successfully. Nov 27 05:01:30 localhost podman[310301]: 2025-11-27 10:01:29.999192349 +0000 UTC m=+0.088611409 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm) Nov 27 05:01:30 localhost podman[310302]: 2025-11-27 10:01:30.014307856 +0000 UTC m=+0.093701165 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:01:30 localhost podman[310302]: 2025-11-27 10:01:30.054087179 +0000 UTC m=+0.133480498 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:01:30 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:01:30 localhost podman[310308]: 2025-11-27 10:01:30.069548175 +0000 UTC m=+0.147190627 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., architecture=x86_64, container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc., name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, vcs-type=git, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., release=1755695350) Nov 27 05:01:30 localhost podman[310308]: 2025-11-27 10:01:30.08607534 +0000 UTC m=+0.163717822 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.openshift.expose-services=, managed_by=edpm_ansible, maintainer=Red Hat, Inc., config_id=edpm, architecture=x86_64, distribution-scope=public, io.buildah.version=1.33.7, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, container_name=openstack_network_exporter) Nov 27 05:01:30 localhost podman[310301]: 2025-11-27 10:01:30.096253255 +0000 UTC m=+0.185672365 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, tcib_managed=true, config_id=edpm, container_name=ceilometer_agent_compute) Nov 27 05:01:30 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:01:30 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:01:30 localhost nova_compute[284026]: 2025-11-27 10:01:30.315 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:30 localhost systemd[1]: tmp-crun.yy6T9o.mount: Deactivated successfully. Nov 27 05:01:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v59: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 17 KiB/s rd, 2.1 MiB/s wr, 23 op/s Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #22. Immutable memtables: 0. Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.379220) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 9] Flushing memtable with next log file: 22 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691379294, "job": 9, "event": "flush_started", "num_memtables": 1, "num_entries": 814, "num_deletes": 256, "total_data_size": 1376617, "memory_usage": 1540096, "flush_reason": "Manual Compaction"} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 9] Level-0 flush table #23: started Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691388353, "cf_name": "default", "job": 9, "event": "table_file_creation", "file_number": 23, "file_size": 898748, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 15463, "largest_seqno": 16272, "table_properties": {"data_size": 895200, "index_size": 1339, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1157, "raw_key_size": 8369, "raw_average_key_size": 19, "raw_value_size": 887774, "raw_average_value_size": 2022, "num_data_blocks": 60, "num_entries": 439, "num_filter_entries": 439, "num_deletions": 256, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237646, "oldest_key_time": 1764237646, "file_creation_time": 1764237691, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 9] Flush lasted 9170 microseconds, and 3685 cpu microseconds. Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.388399) [db/flush_job.cc:967] [default] [JOB 9] Level-0 flush table #23: 898748 bytes OK Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.388422) [db/memtable_list.cc:519] [default] Level-0 commit table #23 started Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.392279) [db/memtable_list.cc:722] [default] Level-0 commit table #23: memtable #1 done Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.392308) EVENT_LOG_v1 {"time_micros": 1764237691392301, "job": 9, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.392331) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 9] Try to delete WAL files size 1372331, prev total WAL file size 1372655, number of live WAL files 2. Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.393030) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6C6F676D0033373638' seq:72057594037927935, type:22 .. '6C6F676D0034303230' seq:0, type:0; will stop at (end) Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 10] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 9 Base level 0, inputs: [23(877KB)], [21(19MB)] Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691393070, "job": 10, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [23], "files_L6": [21], "score": -1, "input_data_size": 20869177, "oldest_snapshot_seqno": -1} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 10] Generated table #24: 11966 keys, 20757470 bytes, temperature: kUnknown Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691568245, "cf_name": "default", "job": 10, "event": "table_file_creation", "file_number": 24, "file_size": 20757470, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 20689770, "index_size": 36687, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 29957, "raw_key_size": 322061, "raw_average_key_size": 26, "raw_value_size": 20486438, "raw_average_value_size": 1712, "num_data_blocks": 1393, "num_entries": 11966, "num_filter_entries": 11966, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237691, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 24, "seqno_to_time_mapping": "N/A"}} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.568653) [db/compaction/compaction_job.cc:1663] [default] [JOB 10] Compacted 1@0 + 1@6 files to L6 => 20757470 bytes Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.570264) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 119.1 rd, 118.4 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.9, 19.0 +0.0 blob) out(19.8 +0.0 blob), read-write-amplify(46.3) write-amplify(23.1) OK, records in: 12501, records dropped: 535 output_compression: NoCompression Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.570294) EVENT_LOG_v1 {"time_micros": 1764237691570280, "job": 10, "event": "compaction_finished", "compaction_time_micros": 175276, "compaction_time_cpu_micros": 52910, "output_level": 6, "num_output_files": 1, "total_output_size": 20757470, "num_input_records": 12501, "num_output_records": 11966, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000023.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691570597, "job": 10, "event": "table_file_deletion", "file_number": 23} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000021.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237691573410, "job": 10, "event": "table_file_deletion", "file_number": 21} Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.392949) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.573590) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.573599) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.573602) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.573606) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:01:31.573609) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:01:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v60: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 16 KiB/s rd, 2.0 MiB/s wr, 23 op/s Nov 27 05:01:33 localhost nova_compute[284026]: 2025-11-27 10:01:33.772 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v61: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 14 KiB/s rd, 1.7 MiB/s wr, 19 op/s Nov 27 05:01:35 localhost nova_compute[284026]: 2025-11-27 10:01:35.317 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v62: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail; 14 KiB/s rd, 1.7 MiB/s wr, 19 op/s Nov 27 05:01:38 localhost podman[242678]: time="2025-11-27T10:01:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:01:38 localhost nova_compute[284026]: 2025-11-27 10:01:38.775 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:38 localhost podman[242678]: @ - - [27/Nov/2025:10:01:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:01:38 localhost podman[242678]: @ - - [27/Nov/2025:10:01:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19718 "" "Go-http-client/1.1" Nov 27 05:01:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v63: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:01:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:01:40 localhost podman[310360]: 2025-11-27 10:01:39.999857642 +0000 UTC m=+0.091256881 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:01:40 localhost podman[310359]: 2025-11-27 10:01:40.033233781 +0000 UTC m=+0.132399198 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 05:01:40 localhost podman[310360]: 2025-11-27 10:01:40.059487118 +0000 UTC m=+0.150886327 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:01:40 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:01:40 localhost podman[310359]: 2025-11-27 10:01:40.104158052 +0000 UTC m=+0.203323399 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:01:40 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:01:40 localhost nova_compute[284026]: 2025-11-27 10:01:40.320 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v64: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v65: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:43.572 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:01:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:43.572 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:01:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:43.573 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:01:43 localhost nova_compute[284026]: 2025-11-27 10:01:43.780 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.137 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.138 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.138 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.142 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ced14236-1ff9-4d9d-86c1-19a8734625b3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.138622', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '140d8abe-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': 'cc3e876a9041457f2141fa52c3581c71abd2000674570c18649839f1be625a5a'}]}, 'timestamp': '2025-11-27 10:01:44.143424', '_unique_id': 'cf488007ed9a40a4a968dfc498956ecd'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.146 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.146 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6566f967-9a61-48e7-8ac4-67c4c4aecc65', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.146681', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '140e1f56-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '3f63d94f1ba431f03587ce3fb018a6b059ada41e7d7ce26911575077eaeffc88'}]}, 'timestamp': '2025-11-27 10:01:44.147220', '_unique_id': 'f3eaaa6947014e25a947489bb8de0085'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.148 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.149 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6f19d8c8-9407-4ef3-9705-4940759148ac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:01:44.149540', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '14112f16-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.354673955, 'message_signature': '1c8ee0b2aaf1a8579160a2927589e5b47f68367b609bd47c7f9a583248e353ce'}]}, 'timestamp': '2025-11-27 10:01:44.167222', '_unique_id': '26237fa2c65c4d6fb3d30bebffb4cfb9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.169 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c1920873-61de-467b-9b52-003e2dc0a5b6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.169648', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1411a0c2-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': 'd28280241763722e58329b76bdef539db68bbd3eaa9b6d1ae57d1300e4302573'}]}, 'timestamp': '2025-11-27 10:01:44.170135', '_unique_id': 'eb60d0fd8bd241c4b36f2966f101b7c0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.172 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.172 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f7d7a0e3-091d-4e60-a6a9-62279175199d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.172387', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '14120d1e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '2634efdfe13339422b3b9fbb64b86c7894f6758eb10dce1722ffd845f138e695'}]}, 'timestamp': '2025-11-27 10:01:44.172978', '_unique_id': 'ca1a9fd9b5764cb2b08d95644b1beffc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.175 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.186 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.187 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'fc30cd2b-535a-4e03-b029-efbf660a7049', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.175227', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '14143c56-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': '33d0822e34933f93afefc879b4114ced608df7af25f551feb4633cad52e1ec73'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.175227', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '14144d5e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': 'cacb4c7461a96be7f979818aaa41e1c2732d3885e42b4267fe63188b66ec826e'}]}, 'timestamp': '2025-11-27 10:01:44.187688', '_unique_id': 'b10fc992ec1d4f63a3f7b482b570b95d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.189 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.190 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd6da6614-8993-4131-99a3-961d62f1b1d2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.190009', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '1414bba4-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '014819d8a6bc801d3b6b7b25dd0fd3e92982eb1e6a06cbe451f024999b0f7c88'}]}, 'timestamp': '2025-11-27 10:01:44.190483', '_unique_id': 'c7418960c1554f0b82cb87b02daf9616'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.192 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.224 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '268cb182-abc7-45ee-a5bf-67576d84238c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.192929', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '1419df8a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '384ed7382042590aaf3f6bd16065fe6c76a3901916aefbcf6961ef44f1246599'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.192929', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '1419f038-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '3db1f431c4a6d41fdd90701bc87013e82d86a7021a67525beea70b2207eb57e0'}]}, 'timestamp': '2025-11-27 10:01:44.224614', '_unique_id': '193fed44babf43c091e1d86ad19495ae'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.226 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1d6a8bba-09c4-4075-99cc-52c8ec2cf4a6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.226870', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141a5afa-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '1081a4444ed47f39ad582624a9341ae5cbfea778a2d91a9fd91eb42185152268'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.226870', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141a6b1c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '803fb7debbf02e2ea709b5af39dea553755754638d00e6a161d832dbb4960d8d'}]}, 'timestamp': '2025-11-27 10:01:44.227753', '_unique_id': '878c1fb2d11c4d5da6fca8a83b8b8903'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.229 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.229 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ddc1cd6e-89da-4400-8e15-54d5fb15ed56', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.229906', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141ad160-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': '21f0012be082553f26248784ccd0daabcc024149780855e714fd790337929941'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.229906', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141ae344-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': '364b5904c1bd848d78548a1f90f90dc528e9a8fe7a03983dd32a4c159d902e9c'}]}, 'timestamp': '2025-11-27 10:01:44.230791', '_unique_id': '0b74bcc79689424caa21e34c77db7177'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e70aad6c-41f6-41cf-a729-cda0d96bc826', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.233316', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '141b586a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '4dcb9eaccc38ccf80a3d907a27b269d1b6a42083ac745632834eae75c11a7ab0'}]}, 'timestamp': '2025-11-27 10:01:44.233817', '_unique_id': 'efc8e94dc2b24166802103c81c6be40a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.235 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.235 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '908c5da6-363f-4ade-9bb2-727d5c9c929b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.235943', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '141bbd28-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '5ae7fe1fd5fb47dc25a1019d67d0607e7d39f8a37ce62d5a5158f52232f8bc89'}]}, 'timestamp': '2025-11-27 10:01:44.236389', '_unique_id': '8d42ced8020a46249728ce28e88a1326'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.238 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.238 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.238 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.238 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8deb431f-e5e8-4c78-9ba5-cda417c39eca', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.238831', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141c2e66-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'af9855bcd2753c199da782b1183f709916762726f8d49ad16d4a521498784509'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.238831', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141c3e2e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '4d4fefea5784cf18823f5a9c1244accf250ee366f82ac114b3e47939063c2fba'}]}, 'timestamp': '2025-11-27 10:01:44.239705', '_unique_id': '019dce20b3af4963a233f9fe58f5d8e9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.241 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.242 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f1e3bb40-4455-41ca-8fa7-f01f99c75b4d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.242125', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141caf58-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': 'd007a394810fdfb8eb473bd14662ddd1554e5d168c6fdd5fa06a824986bdffc7'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.242125', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141cc178-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.363284537, 'message_signature': '3ca5361fd9b80226540012fca31effc984abc712749aad730aef3d565376696f'}]}, 'timestamp': '2025-11-27 10:01:44.243088', '_unique_id': '317703dd77b540e39d9e8afda8bc52eb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e1a0a358-8522-4b8d-bb55-1b2cd365f9ec', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.245432', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141d31f8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'eb56bb9847d03e8380d2b1436b652995acc3c540a55f6ac6dd63eeaccd21a14f'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.245432', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141d4260-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'df399ce4ecd6dc92dfa2b8d862d3ab348a288d8db5e310e792c0c4deb503685c'}]}, 'timestamp': '2025-11-27 10:01:44.246327', '_unique_id': '0a7c3244530d4af985166fc17f5afe86'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd6691a25-897a-447e-afb6-e557ed959284', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.248578', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '141dab4c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '41227f65682147ad859d2a6f42089fab10841e4522f94b0f7b040ff33f6887c5'}]}, 'timestamp': '2025-11-27 10:01:44.249041', '_unique_id': '162ae85cc2d647be80540ab164fb0a50'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.251 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.251 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 14850000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ed8ad24f-de5a-4146-864c-3961e6bd652a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 14850000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:01:44.251188', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '141e1078-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.354673955, 'message_signature': 'a7135b963cc5cbf238ab8fbdfebb343d7fba6cfcdbb93030fff4df7889f5b62b'}]}, 'timestamp': '2025-11-27 10:01:44.251655', '_unique_id': 'de3ba753ea61475e974064d917d5f1e0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e8b4083a-2cd9-4e92-8178-d8e3feb800f8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.253708', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '141e7324-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '8da714e7ff71a323e61c71b2235ae15f4e8fb80b1e70cb807211d091ce2f3e9a'}]}, 'timestamp': '2025-11-27 10:01:44.254155', '_unique_id': '99972f99cd6144c6b127573ea315402e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f32ce4f8-8f3d-4b24-b472-0445c282aa2a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.256169', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141ecfe0-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': '789d63f19db9ef4e5264c39eebc429bf643956c187a8300395088504dcbd4aee'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.256169', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141eda94-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'e2743ffc21f6d31a838bb3742cc25e4cc1eab125e130956cdd379382681ebf86'}]}, 'timestamp': '2025-11-27 10:01:44.256704', '_unique_id': 'e3a2e4190f8b414680d9460f1c74d0b4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.257 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c340f955-af82-4406-a65c-8020c4d4ad24', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:01:44.258000', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '141f1766-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'e7c0f0f84a2a4db9b1faf307e1b6ba3749cda5d3ebf112bfce5ad3bbd06368e8'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:01:44.258000', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '141f2148-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.380984174, 'message_signature': 'f8dfb00f8bfeabc6a8f2081ecef33ecd5b28b2387eca58bcc13ad0c7bad594f7'}]}, 'timestamp': '2025-11-27 10:01:44.258511', '_unique_id': 'd229d859209647a8bf4f7e3dc5c00b5f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7489c0ae-648d-4763-8f52-30af52c5971a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:01:44.259857', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '141f6018-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 11934.32668453, 'message_signature': '1675d200a91404e9d3b73b2d78f144b503a4c852083d894d15c1e399ae780089'}]}, 'timestamp': '2025-11-27 10:01:44.260136', '_unique_id': 'c3bcbc0790174e20a48f6c77d0822c62'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:01:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:01:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:01:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:01:44 Nov 27 05:01:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:01:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:01:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['volumes', 'manila_metadata', '.mgr', 'images', 'vms', 'backups', 'manila_data'] Nov 27 05:01:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:01:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002170138888888889 quantized to 16 (current 16) Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:01:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:01:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:01:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v66: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:45 localhost nova_compute[284026]: 2025-11-27 10:01:45.325 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v67: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:47 localhost ovn_controller[156436]: 2025-11-27T10:01:47Z|00066|memory_trim|INFO|Detected inactivity (last active 30009 ms ago): trimming memory Nov 27 05:01:48 localhost nova_compute[284026]: 2025-11-27 10:01:48.783 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v68: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:50 localhost nova_compute[284026]: 2025-11-27 10:01:50.328 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v69: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:01:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:01:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:52 localhost systemd[1]: tmp-crun.owEc4P.mount: Deactivated successfully. Nov 27 05:01:52 localhost podman[310409]: 2025-11-27 10:01:52.010580769 +0000 UTC m=+0.098406783 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.schema-version=1.0) Nov 27 05:01:52 localhost podman[310409]: 2025-11-27 10:01:52.026913749 +0000 UTC m=+0.114739743 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_id=multipathd, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 05:01:52 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:01:52 localhost systemd[1]: tmp-crun.7gYgC0.mount: Deactivated successfully. Nov 27 05:01:52 localhost podman[310408]: 2025-11-27 10:01:52.115806405 +0000 UTC m=+0.207731749 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent) Nov 27 05:01:52 localhost podman[310408]: 2025-11-27 10:01:52.1271233 +0000 UTC m=+0.219048644 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_managed=true, config_id=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:01:52 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:01:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v70: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 05:01:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 05:01:53 localhost nova_compute[284026]: 2025-11-27 10:01:53.786 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:01:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:01:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:01:54 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:01:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:01:54 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev e125434a-3b5c-41e5-9cbf-df8fd9fef5fa (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:01:54 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev e125434a-3b5c-41e5-9cbf-df8fd9fef5fa (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:01:54 localhost ceph-mgr[290377]: [progress INFO root] Completed event e125434a-3b5c-41e5-9cbf-df8fd9fef5fa (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:01:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:01:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:01:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:54 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:01:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:01:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v71: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:55 localhost nova_compute[284026]: 2025-11-27 10:01:55.331 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:55 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:01:55 localhost openstack_network_exporter[244641]: ERROR 10:01:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:01:55 localhost openstack_network_exporter[244641]: ERROR 10:01:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:01:55 localhost openstack_network_exporter[244641]: ERROR 10:01:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:01:55 localhost openstack_network_exporter[244641]: ERROR 10:01:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:01:55 localhost openstack_network_exporter[244641]: Nov 27 05:01:55 localhost openstack_network_exporter[244641]: ERROR 10:01:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:01:55 localhost openstack_network_exporter[244641]: Nov 27 05:01:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:56.521 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=8, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=7) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:01:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:56.522 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:01:56 localhost nova_compute[284026]: 2025-11-27 10:01:56.522 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:01:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v72: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:58 localhost nova_compute[284026]: 2025-11-27 10:01:58.788 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:01:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v73: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:01:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:01:59.524 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '8'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:02:00 localhost nova_compute[284026]: 2025-11-27 10:02:00.333 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:02:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:02:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:02:01 localhost systemd[1]: tmp-crun.FafMZP.mount: Deactivated successfully. Nov 27 05:02:01 localhost podman[310588]: 2025-11-27 10:02:01.03587668 +0000 UTC m=+0.125129403 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:02:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v74: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:01 localhost podman[310588]: 2025-11-27 10:02:01.049955289 +0000 UTC m=+0.139208062 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:02:01 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:02:01 localhost podman[310589]: 2025-11-27 10:02:01.091730465 +0000 UTC m=+0.181127652 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, maintainer=Red Hat, Inc., managed_by=edpm_ansible, architecture=x86_64, build-date=2025-08-20T13:12:41, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, config_id=edpm, url=https://catalog.redhat.com/en/search?searchType=containers, io.buildah.version=1.33.7, vcs-type=git, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, version=9.6, release=1755695350, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 05:02:01 localhost podman[310589]: 2025-11-27 10:02:01.109827622 +0000 UTC m=+0.199224819 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, name=ubi9-minimal, build-date=2025-08-20T13:12:41, config_id=edpm, vendor=Red Hat, Inc., maintainer=Red Hat, Inc., managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, io.openshift.tags=minimal rhel9, release=1755695350, vcs-type=git, com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, container_name=openstack_network_exporter, version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}) Nov 27 05:02:01 localhost podman[310587]: 2025-11-27 10:02:01.014217056 +0000 UTC m=+0.104244990 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_id=edpm, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:02:01 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:02:01 localhost podman[310587]: 2025-11-27 10:02:01.151427773 +0000 UTC m=+0.241455747 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=edpm, container_name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:02:01 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:02:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v75: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:03 localhost nova_compute[284026]: 2025-11-27 10:02:03.790 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v76: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:05 localhost nova_compute[284026]: 2025-11-27 10:02:05.336 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v77: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:08 localhost podman[242678]: time="2025-11-27T10:02:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:02:08 localhost podman[242678]: @ - - [27/Nov/2025:10:02:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:02:08 localhost nova_compute[284026]: 2025-11-27 10:02:08.793 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:08 localhost podman[242678]: @ - - [27/Nov/2025:10:02:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19717 "" "Go-http-client/1.1" Nov 27 05:02:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v78: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:10 localhost nova_compute[284026]: 2025-11-27 10:02:10.339 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:02:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:02:11 localhost podman[310650]: 2025-11-27 10:02:10.999874286 +0000 UTC m=+0.089569664 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, managed_by=edpm_ansible) Nov 27 05:02:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v79: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:11 localhost podman[310650]: 2025-11-27 10:02:11.085590155 +0000 UTC m=+0.175285523 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:02:11 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:02:11 localhost podman[310651]: 2025-11-27 10:02:11.101312549 +0000 UTC m=+0.175168860 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:02:11 localhost podman[310651]: 2025-11-27 10:02:11.114988048 +0000 UTC m=+0.188844409 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:02:11 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:02:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v80: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:13 localhost nova_compute[284026]: 2025-11-27 10:02:13.796 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v81: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:15 localhost nova_compute[284026]: 2025-11-27 10:02:15.344 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:15 localhost nova_compute[284026]: 2025-11-27 10:02:15.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:16 localhost nova_compute[284026]: 2025-11-27 10:02:16.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v82: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:17 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:02:17.305 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:02:17Z, description=, device_id=fc31abd4-cdc9-4d8f-aa1e-44db2994ab5c, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=13381002-531f-47f7-9aae-9dca234c3d72, ip_allocation=immediate, mac_address=fa:16:3e:2f:9d:91, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=254, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:02:17Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:02:17 localhost podman[310716]: 2025-11-27 10:02:17.541283433 +0000 UTC m=+0.065950278 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:02:17 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:02:17 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:02:17 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:02:17 localhost nova_compute[284026]: 2025-11-27 10:02:17.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:17 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:02:17.875 265123 INFO neutron.agent.dhcp.agent [None req-1915a37e-790f-4ff8-b415-6c34053a88d8 - - - - - -] DHCP configuration for ports {'13381002-531f-47f7-9aae-9dca234c3d72'} is completed#033[00m Nov 27 05:02:18 localhost nova_compute[284026]: 2025-11-27 10:02:18.138 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:18 localhost nova_compute[284026]: 2025-11-27 10:02:18.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:18 localhost nova_compute[284026]: 2025-11-27 10:02:18.826 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v83: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:19 localhost nova_compute[284026]: 2025-11-27 10:02:19.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:19 localhost nova_compute[284026]: 2025-11-27 10:02:19.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:02:20 localhost nova_compute[284026]: 2025-11-27 10:02:20.349 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:20 localhost nova_compute[284026]: 2025-11-27 10:02:20.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v84: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:21 localhost nova_compute[284026]: 2025-11-27 10:02:21.698 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:21 localhost nova_compute[284026]: 2025-11-27 10:02:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.750 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.750 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.751 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:02:22 localhost nova_compute[284026]: 2025-11-27 10:02:22.751 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:02:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:02:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:02:22 localhost systemd[1]: tmp-crun.zSb0ts.mount: Deactivated successfully. Nov 27 05:02:23 localhost podman[310757]: 2025-11-27 10:02:23.005261758 +0000 UTC m=+0.096280435 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:02:23 localhost podman[310757]: 2025-11-27 10:02:23.038982847 +0000 UTC m=+0.130001504 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, container_name=ovn_metadata_agent) Nov 27 05:02:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v85: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:23 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:02:23 localhost podman[310758]: 2025-11-27 10:02:23.099427876 +0000 UTC m=+0.187385570 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:02:23 localhost podman[310758]: 2025-11-27 10:02:23.140012469 +0000 UTC m=+0.227970143 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:02:23 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:02:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:02:23 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/811314525' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.240 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.489s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.286 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.286 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.456 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.457 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11398MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.457 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.457 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.813 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.813 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.814 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:02:23 localhost nova_compute[284026]: 2025-11-27 10:02:23.829 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.062 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.625 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.626 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.653 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.677 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 05:02:24 localhost nova_compute[284026]: 2025-11-27 10:02:24.730 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:02:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v86: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:02:25 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/957354338' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.205 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.475s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.212 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.230 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.233 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.233 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.776s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:02:25 localhost nova_compute[284026]: 2025-11-27 10:02:25.353 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:25 localhost openstack_network_exporter[244641]: ERROR 10:02:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:02:25 localhost openstack_network_exporter[244641]: ERROR 10:02:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:02:25 localhost openstack_network_exporter[244641]: ERROR 10:02:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:02:25 localhost openstack_network_exporter[244641]: ERROR 10:02:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:02:25 localhost openstack_network_exporter[244641]: Nov 27 05:02:25 localhost openstack_network_exporter[244641]: ERROR 10:02:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:02:25 localhost openstack_network_exporter[244641]: Nov 27 05:02:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:02:26.429 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:02:26Z, description=, device_id=7b75881e-3fdb-408f-b7d4-19f6c550a573, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=9e03d189-9257-4e84-a868-04d1b6c4304d, ip_allocation=immediate, mac_address=fa:16:3e:a3:72:9a, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=337, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:02:26Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:02:26 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:02:26 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:02:26 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:02:26 localhost podman[310833]: 2025-11-27 10:02:26.681631904 +0000 UTC m=+0.059485385 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.856 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.856 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.857 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:02:26 localhost nova_compute[284026]: 2025-11-27 10:02:26.857 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:02:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:02:26.858 265123 INFO neutron.agent.dhcp.agent [None req-35d092c0-43e8-44a7-b378-f18405954ee3 - - - - - -] DHCP configuration for ports {'9e03d189-9257-4e84-a868-04d1b6c4304d'} is completed#033[00m Nov 27 05:02:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v87: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:27 localhost nova_compute[284026]: 2025-11-27 10:02:27.328 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:27 localhost nova_compute[284026]: 2025-11-27 10:02:27.595 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:02:27 localhost nova_compute[284026]: 2025-11-27 10:02:27.678 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:02:27 localhost nova_compute[284026]: 2025-11-27 10:02:27.679 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:02:27 localhost nova_compute[284026]: 2025-11-27 10:02:27.680 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:28 localhost nova_compute[284026]: 2025-11-27 10:02:28.808 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:28 localhost nova_compute[284026]: 2025-11-27 10:02:28.809 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 05:02:28 localhost nova_compute[284026]: 2025-11-27 10:02:28.867 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v88: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:30 localhost nova_compute[284026]: 2025-11-27 10:02:30.172 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:02:30 localhost nova_compute[284026]: 2025-11-27 10:02:30.173 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 05:02:30 localhost nova_compute[284026]: 2025-11-27 10:02:30.356 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v89: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:31 localhost nova_compute[284026]: 2025-11-27 10:02:31.201 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 05:02:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:02:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:02:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:02:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:32 localhost systemd[1]: tmp-crun.xceexV.mount: Deactivated successfully. Nov 27 05:02:32 localhost podman[310856]: 2025-11-27 10:02:32.009007878 +0000 UTC m=+0.099450111 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:02:32 localhost podman[310856]: 2025-11-27 10:02:32.021837944 +0000 UTC m=+0.112280137 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:02:32 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:02:32 localhost podman[310855]: 2025-11-27 10:02:32.116836794 +0000 UTC m=+0.208668464 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=edpm, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:02:32 localhost podman[310855]: 2025-11-27 10:02:32.154118568 +0000 UTC m=+0.245950258 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, container_name=ceilometer_agent_compute) Nov 27 05:02:32 localhost systemd[1]: tmp-crun.qzh7cE.mount: Deactivated successfully. Nov 27 05:02:32 localhost podman[310857]: 2025-11-27 10:02:32.169727419 +0000 UTC m=+0.252101004 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, architecture=x86_64, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, io.buildah.version=1.33.7, managed_by=edpm_ansible, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, com.redhat.component=ubi9-minimal-container, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, distribution-scope=public, maintainer=Red Hat, Inc., vcs-type=git) Nov 27 05:02:32 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:02:32 localhost podman[310857]: 2025-11-27 10:02:32.180768427 +0000 UTC m=+0.263141982 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, build-date=2025-08-20T13:12:41, release=1755695350, vcs-type=git, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, name=ubi9-minimal, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, distribution-scope=public, io.openshift.expose-services=, maintainer=Red Hat, Inc., vendor=Red Hat, Inc., version=9.6, config_id=edpm, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container) Nov 27 05:02:32 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:02:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v90: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:33 localhost nova_compute[284026]: 2025-11-27 10:02:33.869 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v91: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:35 localhost nova_compute[284026]: 2025-11-27 10:02:35.358 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v92: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:38 localhost podman[242678]: time="2025-11-27T10:02:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:02:38 localhost podman[242678]: @ - - [27/Nov/2025:10:02:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:02:38 localhost podman[242678]: @ - - [27/Nov/2025:10:02:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19714 "" "Go-http-client/1.1" Nov 27 05:02:38 localhost nova_compute[284026]: 2025-11-27 10:02:38.911 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v93: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:40 localhost nova_compute[284026]: 2025-11-27 10:02:40.362 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v94: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:02:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:02:41 localhost podman[310912]: 2025-11-27 10:02:41.981998836 +0000 UTC m=+0.079039831 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:02:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:42 localhost podman[310912]: 2025-11-27 10:02:42.018326075 +0000 UTC m=+0.115367120 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:02:42 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:02:42 localhost podman[310913]: 2025-11-27 10:02:42.039143306 +0000 UTC m=+0.133777636 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:02:42 localhost podman[310913]: 2025-11-27 10:02:42.04708597 +0000 UTC m=+0.141720280 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:02:42 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:02:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v95: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:02:43.572 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:02:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:02:43.573 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:02:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:02:43.575 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:02:43 localhost nova_compute[284026]: 2025-11-27 10:02:43.939 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:02:44 Nov 27 05:02:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:02:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:02:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['manila_metadata', '.mgr', 'vms', 'images', 'manila_data', 'backups', 'volumes'] Nov 27 05:02:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:02:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002170138888888889 quantized to 16 (current 16) Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:02:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:02:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:02:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v96: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:45 localhost nova_compute[284026]: 2025-11-27 10:02:45.365 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v97: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:48 localhost nova_compute[284026]: 2025-11-27 10:02:48.968 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v98: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:50 localhost nova_compute[284026]: 2025-11-27 10:02:50.367 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v99: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v100: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:02:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:02:53 localhost nova_compute[284026]: 2025-11-27 10:02:53.993 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:54 localhost systemd[1]: tmp-crun.JG2CIX.mount: Deactivated successfully. Nov 27 05:02:54 localhost podman[310960]: 2025-11-27 10:02:54.058434733 +0000 UTC m=+0.148999677 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:02:54 localhost podman[310959]: 2025-11-27 10:02:54.019301623 +0000 UTC m=+0.112335084 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:02:54 localhost podman[310959]: 2025-11-27 10:02:54.102191168 +0000 UTC m=+0.195224579 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:02:54 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:02:54 localhost podman[310960]: 2025-11-27 10:02:54.152746898 +0000 UTC m=+0.243311792 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:02:54 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:02:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v101: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:55 localhost nova_compute[284026]: 2025-11-27 10:02:55.369 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:02:55 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:02:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:02:55 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:02:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:02:55 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 9a3a2254-acf8-4423-80ef-0048c82a64a9 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:02:55 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 9a3a2254-acf8-4423-80ef-0048c82a64a9 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:02:55 localhost ceph-mgr[290377]: [progress INFO root] Completed event 9a3a2254-acf8-4423-80ef-0048c82a64a9 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:02:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:02:55 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:02:55 localhost openstack_network_exporter[244641]: ERROR 10:02:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:02:55 localhost openstack_network_exporter[244641]: ERROR 10:02:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:02:55 localhost openstack_network_exporter[244641]: ERROR 10:02:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:02:55 localhost openstack_network_exporter[244641]: ERROR 10:02:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:02:55 localhost openstack_network_exporter[244641]: Nov 27 05:02:55 localhost openstack_network_exporter[244641]: ERROR 10:02:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:02:55 localhost openstack_network_exporter[244641]: Nov 27 05:02:56 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:02:56 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:02:56 localhost nova_compute[284026]: 2025-11-27 10:02:56.589 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:02:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v102: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:02:58.371 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=9, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=8) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:02:58 localhost nova_compute[284026]: 2025-11-27 10:02:58.371 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:02:58.373 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:02:59 localhost nova_compute[284026]: 2025-11-27 10:02:59.028 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:02:59 localhost neutron_sriov_agent[258162]: 2025-11-27 10:02:59.040 2 INFO neutron.agent.securitygroups_rpc [None req-a4835aec-cc08-4665-b5bf-6ab6ecc6db39 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Security group member updated ['d5a4dbad-2a9b-45ce-81ca-0430479e3c01']#033[00m Nov 27 05:02:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v103: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:02:59 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:02:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:03:00 localhost nova_compute[284026]: 2025-11-27 10:03:00.372 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:03:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v104: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:01 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:01.375 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '9'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #25. Immutable memtables: 0. Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.639993) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 11] Flushing memtable with next log file: 25 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782640046, "job": 11, "event": "flush_started", "num_memtables": 1, "num_entries": 1330, "num_deletes": 251, "total_data_size": 1692901, "memory_usage": 1721760, "flush_reason": "Manual Compaction"} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 11] Level-0 flush table #26: started Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782651290, "cf_name": "default", "job": 11, "event": "table_file_creation", "file_number": 26, "file_size": 1078814, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 16277, "largest_seqno": 17602, "table_properties": {"data_size": 1073480, "index_size": 2805, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1477, "raw_key_size": 11994, "raw_average_key_size": 20, "raw_value_size": 1062674, "raw_average_value_size": 1841, "num_data_blocks": 119, "num_entries": 577, "num_filter_entries": 577, "num_deletions": 251, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237691, "oldest_key_time": 1764237691, "file_creation_time": 1764237782, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 26, "seqno_to_time_mapping": "N/A"}} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 11] Flush lasted 11365 microseconds, and 4591 cpu microseconds. Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.651355) [db/flush_job.cc:967] [default] [JOB 11] Level-0 flush table #26: 1078814 bytes OK Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.651384) [db/memtable_list.cc:519] [default] Level-0 commit table #26 started Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.653946) [db/memtable_list.cc:722] [default] Level-0 commit table #26: memtable #1 done Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.653974) EVENT_LOG_v1 {"time_micros": 1764237782653967, "job": 11, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.653997) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 11] Try to delete WAL files size 1686556, prev total WAL file size 1686556, number of live WAL files 2. Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000022.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.655011) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003131353436' seq:72057594037927935, type:22 .. '7061786F73003131373938' seq:0, type:0; will stop at (end) Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 12] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 11 Base level 0, inputs: [26(1053KB)], [24(19MB)] Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782655065, "job": 12, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [26], "files_L6": [24], "score": -1, "input_data_size": 21836284, "oldest_snapshot_seqno": -1} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 12] Generated table #27: 12018 keys, 19584526 bytes, temperature: kUnknown Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782815057, "cf_name": "default", "job": 12, "event": "table_file_creation", "file_number": 27, "file_size": 19584526, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19517322, "index_size": 36063, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 30085, "raw_key_size": 323628, "raw_average_key_size": 26, "raw_value_size": 19313863, "raw_average_value_size": 1607, "num_data_blocks": 1364, "num_entries": 12018, "num_filter_entries": 12018, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237782, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 27, "seqno_to_time_mapping": "N/A"}} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.815395) [db/compaction/compaction_job.cc:1663] [default] [JOB 12] Compacted 1@0 + 1@6 files to L6 => 19584526 bytes Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.817295) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 136.4 rd, 122.3 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(1.0, 19.8 +0.0 blob) out(18.7 +0.0 blob), read-write-amplify(38.4) write-amplify(18.2) OK, records in: 12543, records dropped: 525 output_compression: NoCompression Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.817326) EVENT_LOG_v1 {"time_micros": 1764237782817313, "job": 12, "event": "compaction_finished", "compaction_time_micros": 160080, "compaction_time_cpu_micros": 51730, "output_level": 6, "num_output_files": 1, "total_output_size": 19584526, "num_input_records": 12543, "num_output_records": 12018, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000026.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782817635, "job": 12, "event": "table_file_deletion", "file_number": 26} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000024.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237782820552, "job": 12, "event": "table_file_deletion", "file_number": 24} Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.654887) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.820709) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.820717) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.820720) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.820723) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:03:02.820726) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:03:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:03:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:03:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:03:03 localhost podman[311081]: 2025-11-27 10:03:02.988065502 +0000 UTC m=+0.078159368 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute) Nov 27 05:03:03 localhost systemd[1]: tmp-crun.QB2oxA.mount: Deactivated successfully. Nov 27 05:03:03 localhost podman[311082]: 2025-11-27 10:03:03.053151555 +0000 UTC m=+0.139589952 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:03:03 localhost podman[311082]: 2025-11-27 10:03:03.067027011 +0000 UTC m=+0.153465378 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:03:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v105: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:03 localhost podman[311083]: 2025-11-27 10:03:03.022576277 +0000 UTC m=+0.103747462 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-type=git, com.redhat.component=ubi9-minimal-container, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc., name=ubi9-minimal, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, io.buildah.version=1.33.7, vendor=Red Hat, Inc., release=1755695350, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, distribution-scope=public, container_name=openstack_network_exporter, managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI) Nov 27 05:03:03 localhost podman[311083]: 2025-11-27 10:03:03.105802201 +0000 UTC m=+0.186973366 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.component=ubi9-minimal-container, url=https://catalog.redhat.com/en/search?searchType=containers, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, managed_by=edpm_ansible, io.openshift.tags=minimal rhel9, build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., config_id=edpm, vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., release=1755695350, io.openshift.expose-services=, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, name=ubi9-minimal, architecture=x86_64, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public) Nov 27 05:03:03 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:03:03 localhost podman[311081]: 2025-11-27 10:03:03.123311115 +0000 UTC m=+0.213404931 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:03:03 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:03:03 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:03:04 localhost nova_compute[284026]: 2025-11-27 10:03:04.031 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v106: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:05 localhost neutron_sriov_agent[258162]: 2025-11-27 10:03:05.317 2 INFO neutron.agent.securitygroups_rpc [None req-3a756a3d-ce49-4a08-a782-63066c72d252 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Security group member updated ['d5a4dbad-2a9b-45ce-81ca-0430479e3c01']#033[00m Nov 27 05:03:05 localhost nova_compute[284026]: 2025-11-27 10:03:05.374 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:05 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:05.838 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:05Z, description=, device_id=a0c4d84a-974c-45d9-84bb-83ff16c47e9f, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=0b3b0f26-49c3-4a12-b88c-ff845da4b9e4, ip_allocation=immediate, mac_address=fa:16:3e:99:82:34, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=442, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:03:05Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:03:06 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:03:06 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:06 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:06 localhost podman[311161]: 2025-11-27 10:03:06.205780183 +0000 UTC m=+0.067005667 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:03:06 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:06.601 265123 INFO neutron.agent.dhcp.agent [None req-e6878d45-0798-49d2-a620-348603a3298d - - - - - -] DHCP configuration for ports {'0b3b0f26-49c3-4a12-b88c-ff845da4b9e4'} is completed#033[00m Nov 27 05:03:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v107: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:08 localhost nova_compute[284026]: 2025-11-27 10:03:08.336 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:08 localhost podman[242678]: time="2025-11-27T10:03:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:03:08 localhost podman[242678]: @ - - [27/Nov/2025:10:03:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:03:08 localhost podman[242678]: @ - - [27/Nov/2025:10:03:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19723 "" "Go-http-client/1.1" Nov 27 05:03:09 localhost nova_compute[284026]: 2025-11-27 10:03:09.069 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v108: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:10 localhost nova_compute[284026]: 2025-11-27 10:03:10.377 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v109: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:11 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:11.986 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:10Z, description=, device_id=b64b82a5-274f-4e19-a72b-e17e014de7b0, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e6274c53-2a35-4b2d-8ada-98687d92eee5, ip_allocation=immediate, mac_address=fa:16:3e:4c:9b:c0, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=471, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:03:11Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:03:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:12 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:03:12 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:12 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:12 localhost podman[311197]: 2025-11-27 10:03:12.586789226 +0000 UTC m=+0.066787860 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:03:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:03:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:03:12 localhost podman[311212]: 2025-11-27 10:03:12.708224956 +0000 UTC m=+0.089334181 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:03:12 localhost systemd[1]: tmp-crun.Dc35k4.mount: Deactivated successfully. Nov 27 05:03:12 localhost podman[311211]: 2025-11-27 10:03:12.773137884 +0000 UTC m=+0.158410212 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:03:12 localhost podman[311212]: 2025-11-27 10:03:12.799083747 +0000 UTC m=+0.180193032 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:03:12 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:03:12 localhost podman[311211]: 2025-11-27 10:03:12.843220602 +0000 UTC m=+0.228492890 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:03:12 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:03:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v110: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:13.200 265123 INFO neutron.agent.dhcp.agent [None req-d26e0114-5855-41b6-9978-cdf3df661fe7 - - - - - -] DHCP configuration for ports {'e6274c53-2a35-4b2d-8ada-98687d92eee5'} is completed#033[00m Nov 27 05:03:14 localhost nova_compute[284026]: 2025-11-27 10:03:14.071 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v111: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:15 localhost nova_compute[284026]: 2025-11-27 10:03:15.379 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:16 localhost nova_compute[284026]: 2025-11-27 10:03:16.255 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:16 localhost nova_compute[284026]: 2025-11-27 10:03:16.424 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v112: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:18 localhost nova_compute[284026]: 2025-11-27 10:03:18.761 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:18 localhost nova_compute[284026]: 2025-11-27 10:03:18.761 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:18 localhost nova_compute[284026]: 2025-11-27 10:03:18.762 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v113: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:19 localhost nova_compute[284026]: 2025-11-27 10:03:19.097 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:19 localhost nova_compute[284026]: 2025-11-27 10:03:19.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:20 localhost nova_compute[284026]: 2025-11-27 10:03:20.384 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:20 localhost nova_compute[284026]: 2025-11-27 10:03:20.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:20 localhost nova_compute[284026]: 2025-11-27 10:03:20.729 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v114: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:21 localhost nova_compute[284026]: 2025-11-27 10:03:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:21 localhost nova_compute[284026]: 2025-11-27 10:03:21.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:03:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:22 localhost nova_compute[284026]: 2025-11-27 10:03:22.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v115: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.128 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.808 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.809 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.809 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.810 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:03:24 localhost nova_compute[284026]: 2025-11-27 10:03:24.810 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:03:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:03:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:03:25 localhost systemd[1]: tmp-crun.oZtxMa.mount: Deactivated successfully. Nov 27 05:03:25 localhost podman[311268]: 2025-11-27 10:03:25.00635605 +0000 UTC m=+0.094711856 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_id=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:03:25 localhost podman[311268]: 2025-11-27 10:03:25.020945965 +0000 UTC m=+0.109301781 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:03:25 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:03:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v116: 177 pgs: 177 active+clean; 145 MiB data, 724 MiB used, 41 GiB / 42 GiB avail Nov 27 05:03:25 localhost podman[311267]: 2025-11-27 10:03:25.097729725 +0000 UTC m=+0.191680793 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent) Nov 27 05:03:25 localhost podman[311267]: 2025-11-27 10:03:25.129015362 +0000 UTC m=+0.222966460 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:03:25 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:03:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:03:25 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1181691291' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.366 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.556s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.390 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.606 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.607 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:03:25 localhost openstack_network_exporter[244641]: ERROR 10:03:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:03:25 localhost openstack_network_exporter[244641]: ERROR 10:03:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:03:25 localhost openstack_network_exporter[244641]: ERROR 10:03:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:03:25 localhost openstack_network_exporter[244641]: ERROR 10:03:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:03:25 localhost openstack_network_exporter[244641]: Nov 27 05:03:25 localhost openstack_network_exporter[244641]: ERROR 10:03:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:03:25 localhost openstack_network_exporter[244641]: Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.850 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.852 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11384MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.853 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:03:25 localhost nova_compute[284026]: 2025-11-27 10:03:25.853 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:03:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v117: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 33 op/s Nov 27 05:03:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:27 localhost nova_compute[284026]: 2025-11-27 10:03:27.826 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:03:27 localhost nova_compute[284026]: 2025-11-27 10:03:27.827 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:03:27 localhost nova_compute[284026]: 2025-11-27 10:03:27.828 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:03:27 localhost nova_compute[284026]: 2025-11-27 10:03:27.930 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:03:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:03:28 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/66746130' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:03:28 localhost nova_compute[284026]: 2025-11-27 10:03:28.375 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.444s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:03:28 localhost nova_compute[284026]: 2025-11-27 10:03:28.382 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:03:28 localhost nova_compute[284026]: 2025-11-27 10:03:28.409 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:03:28 localhost nova_compute[284026]: 2025-11-27 10:03:28.412 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:03:28 localhost nova_compute[284026]: 2025-11-27 10:03:28.412 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.559s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:03:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v118: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 33 op/s Nov 27 05:03:29 localhost nova_compute[284026]: 2025-11-27 10:03:29.162 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:30 localhost nova_compute[284026]: 2025-11-27 10:03:30.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:30 localhost nova_compute[284026]: 2025-11-27 10:03:30.413 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:03:30 localhost nova_compute[284026]: 2025-11-27 10:03:30.414 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:03:30 localhost nova_compute[284026]: 2025-11-27 10:03:30.414 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:03:30 localhost nova_compute[284026]: 2025-11-27 10:03:30.860 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v119: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 33 op/s Nov 27 05:03:31 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:03:31 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:31 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:31 localhost podman[311364]: 2025-11-27 10:03:31.141755822 +0000 UTC m=+0.063055710 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:03:31 localhost nova_compute[284026]: 2025-11-27 10:03:31.263 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:03:31 localhost nova_compute[284026]: 2025-11-27 10:03:31.263 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:03:31 localhost nova_compute[284026]: 2025-11-27 10:03:31.264 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:03:31 localhost nova_compute[284026]: 2025-11-27 10:03:31.264 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:03:31 localhost ovn_controller[156436]: 2025-11-27T10:03:31Z|00067|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:03:31 localhost nova_compute[284026]: 2025-11-27 10:03:31.405 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:32 localhost nova_compute[284026]: 2025-11-27 10:03:32.652 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:03:32 localhost nova_compute[284026]: 2025-11-27 10:03:32.681 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:03:32 localhost nova_compute[284026]: 2025-11-27 10:03:32.682 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:03:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v120: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 33 op/s Nov 27 05:03:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:03:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:03:33 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:03:34 localhost podman[311386]: 2025-11-27 10:03:34.00686513 +0000 UTC m=+0.095622862 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:03:34 localhost podman[311386]: 2025-11-27 10:03:34.042806032 +0000 UTC m=+0.131563774 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:03:34 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:03:34 localhost podman[311388]: 2025-11-27 10:03:34.062765914 +0000 UTC m=+0.147252170 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, managed_by=edpm_ansible, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., maintainer=Red Hat, Inc., name=ubi9-minimal, version=9.6, build-date=2025-08-20T13:12:41, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., vcs-type=git, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., container_name=openstack_network_exporter, io.openshift.tags=minimal rhel9, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64) Nov 27 05:03:34 localhost podman[311388]: 2025-11-27 10:03:34.100926587 +0000 UTC m=+0.185412903 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, io.buildah.version=1.33.7, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, config_id=edpm, io.openshift.tags=minimal rhel9, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=9.6, build-date=2025-08-20T13:12:41, io.openshift.expose-services=, distribution-scope=public, maintainer=Red Hat, Inc., architecture=x86_64, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, managed_by=edpm_ansible) Nov 27 05:03:34 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:03:34 localhost podman[311387]: 2025-11-27 10:03:34.12578694 +0000 UTC m=+0.212077996 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:03:34 localhost podman[311387]: 2025-11-27 10:03:34.159031481 +0000 UTC m=+0.245322497 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:03:34 localhost nova_compute[284026]: 2025-11-27 10:03:34.201 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:34 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:03:35 localhost systemd[1]: tmp-crun.KW3pej.mount: Deactivated successfully. Nov 27 05:03:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v121: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 33 op/s Nov 27 05:03:35 localhost nova_compute[284026]: 2025-11-27 10:03:35.396 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v122: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.7 MiB/s rd, 1.8 MiB/s wr, 43 op/s Nov 27 05:03:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:38 localhost podman[242678]: time="2025-11-27T10:03:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:03:38 localhost podman[242678]: @ - - [27/Nov/2025:10:03:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:03:38 localhost podman[242678]: @ - - [27/Nov/2025:10:03:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19722 "" "Go-http-client/1.1" Nov 27 05:03:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v123: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 7.2 KiB/s rd, 12 KiB/s wr, 9 op/s Nov 27 05:03:39 localhost nova_compute[284026]: 2025-11-27 10:03:39.202 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:40 localhost nova_compute[284026]: 2025-11-27 10:03:40.399 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:41 localhost nova_compute[284026]: 2025-11-27 10:03:41.072 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:41 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:41.072 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=10, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=9) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:03:41 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:41.074 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 7 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:03:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v124: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 12 KiB/s wr, 73 op/s Nov 27 05:03:41 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:41.987 265123 INFO neutron.agent.linux.ip_lib [None req-b7810261-5931-4ae5-9078-1652f1a00d4d - - - - - -] Device tapae33bdba-c0 cannot be used as it has no MAC address#033[00m Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.011 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost kernel: device tapae33bdba-c0 entered promiscuous mode Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.035 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost ovn_controller[156436]: 2025-11-27T10:03:42Z|00068|binding|INFO|Claiming lport ae33bdba-c03c-4d06-a040-bfb31931b448 for this chassis. Nov 27 05:03:42 localhost ovn_controller[156436]: 2025-11-27T10:03:42Z|00069|binding|INFO|ae33bdba-c03c-4d06-a040-bfb31931b448: Claiming unknown Nov 27 05:03:42 localhost NetworkManager[5971]: [1764237822.0418] manager: (tapae33bdba-c0): new Generic device (/org/freedesktop/NetworkManager/Devices/19) Nov 27 05:03:42 localhost systemd-udevd[311458]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.078 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost ovn_controller[156436]: 2025-11-27T10:03:42Z|00070|binding|INFO|Setting lport ae33bdba-c03c-4d06-a040-bfb31931b448 ovn-installed in OVS Nov 27 05:03:42 localhost ovn_controller[156436]: 2025-11-27T10:03:42Z|00071|binding|INFO|Setting lport ae33bdba-c03c-4d06-a040-bfb31931b448 up in Southbound Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.084 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:42.082 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.3/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-75d79a30-b416-43bb-83b3-cd22cce0e128', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-75d79a30-b416-43bb-83b3-cd22cce0e128', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '2f8a8262d8074ffd876910fbc25620fe', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=6fde6c61-2676-4137-9e99-778e99b077f8, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=ae33bdba-c03c-4d06-a040-bfb31931b448) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:03:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:42.085 162092 INFO neutron.agent.ovn.metadata.agent [-] Port ae33bdba-c03c-4d06-a040-bfb31931b448 in datapath 75d79a30-b416-43bb-83b3-cd22cce0e128 bound to our chassis#033[00m Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.087 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:42.090 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 59aae138-a2bc-47ad-8d26-c54af147aac5 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:03:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:42.091 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 75d79a30-b416-43bb-83b3-cd22cce0e128, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:03:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:42.096 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[77621225-bc0e-4840-95cb-88dfeb778dfe]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.130 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:42.136 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:41Z, description=, device_id=36c94130-e961-4421-bbd2-0a6d6b2fa5ea, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=a6fd71db-1e24-4068-9b73-f1eb055922d4, ip_allocation=immediate, mac_address=fa:16:3e:8c:22:40, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=579, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:03:41Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:03:42 localhost nova_compute[284026]: 2025-11-27 10:03:42.163 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:42 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:03:42 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:42 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:42 localhost podman[311488]: 2025-11-27 10:03:42.408698965 +0000 UTC m=+0.070556682 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:03:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:42.597 265123 INFO neutron.agent.dhcp.agent [None req-4fad70b9-ffdb-47fd-9546-3747fe406a3e - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:42Z, description=, device_id=ef2bcfd0-753b-4017-b94e-fb3e2a8510cb, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=d7b115cb-ff65-4618-b2db-57a37d6ff7e0, ip_allocation=immediate, mac_address=fa:16:3e:49:24:a7, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=580, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:03:42Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:03:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:42.783 265123 INFO neutron.agent.dhcp.agent [None req-09870d1c-0240-4803-a744-8dfedfd5590f - - - - - -] DHCP configuration for ports {'a6fd71db-1e24-4068-9b73-f1eb055922d4'} is completed#033[00m Nov 27 05:03:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:03:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:03:42 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 6 addresses Nov 27 05:03:42 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:42 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:42 localhost podman[311541]: 2025-11-27 10:03:42.963863173 +0000 UTC m=+0.084186371 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:03:43 localhost systemd[1]: tmp-crun.u4z8Ce.mount: Deactivated successfully. Nov 27 05:03:43 localhost podman[311552]: 2025-11-27 10:03:43.032920094 +0000 UTC m=+0.112287764 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:03:43 localhost podman[311552]: 2025-11-27 10:03:43.046063429 +0000 UTC m=+0.125431029 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:03:43 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:03:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v125: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 12 KiB/s wr, 73 op/s Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.122 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Creating tmpfile /var/lib/nova/instances/tmp110wzs4o to notify to other compute nodes that they should mount the same storage. _create_shared_storage_test_file /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10041#033[00m Nov 27 05:03:43 localhost podman[311551]: 2025-11-27 10:03:43.14094726 +0000 UTC m=+0.230271609 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, container_name=ovn_controller) Nov 27 05:03:43 localhost podman[311551]: 2025-11-27 10:03:43.225238773 +0000 UTC m=+0.314563182 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=ovn_controller, managed_by=edpm_ansible) Nov 27 05:03:43 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.291 284030 DEBUG nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] destination check data is LibvirtLiveMigrateData(bdms=,block_migration=,disk_available_mb=13312,disk_over_commit=,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp110wzs4o',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path=,is_shared_block_storage=,is_shared_instance_path=,is_volume_backed=,migration=,old_vol_attachment_ids=,serial_listen_addr=None,serial_listen_ports=,src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=,target_connect_addr=,vifs=[VIFMigrateData],wait_for_vif_plugged=) check_can_live_migrate_destination /usr/lib/python3.9/site-packages/nova/compute/manager.py:8476#033[00m Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.314 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquiring lock "compute-rpcapi-router" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.316 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquired lock "compute-rpcapi-router" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:03:43 localhost podman[311628]: Nov 27 05:03:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:43.322 265123 INFO neutron.agent.dhcp.agent [None req-d79b3f1f-3aa0-47cd-99f9-302c0b70488c - - - - - -] DHCP configuration for ports {'d7b115cb-ff65-4618-b2db-57a37d6ff7e0'} is completed#033[00m Nov 27 05:03:43 localhost podman[311628]: 2025-11-27 10:03:43.33107176 +0000 UTC m=+0.132976833 container create 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.347 284030 INFO nova.compute.rpcapi [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Automatically selected compute RPC version 6.2 from minimum service version 66#033[00m Nov 27 05:03:43 localhost nova_compute[284026]: 2025-11-27 10:03:43.349 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Releasing lock "compute-rpcapi-router" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:03:43 localhost systemd[1]: Started libpod-conmon-51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f.scope. Nov 27 05:03:43 localhost podman[311628]: 2025-11-27 10:03:43.286294937 +0000 UTC m=+0.088200080 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:03:43 localhost systemd[1]: Started libcrun container. Nov 27 05:03:43 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/14f73f1e31358fea30c828fc2907fbab9fa54e1cddf6e2ea4115fbbf651e33c4/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:03:43 localhost podman[311628]: 2025-11-27 10:03:43.40639942 +0000 UTC m=+0.208304493 container init 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:03:43 localhost podman[311628]: 2025-11-27 10:03:43.414479329 +0000 UTC m=+0.216384402 container start 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:03:43 localhost dnsmasq[311647]: started, version 2.85 cachesize 150 Nov 27 05:03:43 localhost dnsmasq[311647]: DNS service limited to local subnets Nov 27 05:03:43 localhost dnsmasq[311647]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:03:43 localhost dnsmasq[311647]: warning: no upstream servers configured Nov 27 05:03:43 localhost dnsmasq-dhcp[311647]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:03:43 localhost dnsmasq[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/addn_hosts - 0 addresses Nov 27 05:03:43 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/host Nov 27 05:03:43 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/opts Nov 27 05:03:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:43.572 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:03:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:43.573 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:03:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:43.574 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:03:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:43.576 265123 INFO neutron.agent.dhcp.agent [None req-f793b713-a5f7-4399-bccd-a66ce67fb32a - - - - - -] DHCP configuration for ports {'bb03a32f-b17a-4494-a02c-dae0d84f9d49'} is completed#033[00m Nov 27 05:03:44 localhost nova_compute[284026]: 2025-11-27 10:03:44.087 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.136 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.137 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.137 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.141 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b3dfcf70-bfc4-4341-b52d-33a5aac48dcf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.137966', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b93e0b8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '01361e5ae0b1b8a8e8671e4b6b4a93b83016eae937c539b651610bbcedaba196'}]}, 'timestamp': '2025-11-27 10:03:44.142094', '_unique_id': 'f5cc0cb804a24c9e8e41d6166f315fc5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.143 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.144 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.145 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e43d13d5-1090-4ce4-9c8c-591bc84cf8cc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.145040', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b946ad8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '5aa3db1ffc09e3786258424f15eb1652f51ca9611eead3f49f77990e87c6615c'}]}, 'timestamp': '2025-11-27 10:03:44.145546', '_unique_id': 'b66ebdab216545e59215babdc89d5e1d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.146 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.147 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.147 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'afb4a462-dd3c-416b-835e-d2a969478668', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.147914', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b94db12-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '3c279e9f6134c9e9fcc0694985853e09932be4d7e8fa5cd4bca536524257ba6f'}]}, 'timestamp': '2025-11-27 10:03:44.148385', '_unique_id': '9616369448ca4a9a903e534ee9b38203'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.150 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 15450000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f06f4915-ccc2-437c-b90b-804d55f041b3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 15450000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:03:44.150585', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '5b97b24c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.354343447, 'message_signature': '4865882d8ef30ebc323f7382d91d536058f5b332b51b69fd4411f61551988aff'}]}, 'timestamp': '2025-11-27 10:03:44.166996', '_unique_id': 'e30c1ee9ffc54c229e98083ac8429022'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.169 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'dfd05af6-c91f-4e3d-8c74-31184f6df8bf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.169188', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b9819f8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': 'd4ec4778ceb151c1fe45c440aa63bdac0761c36d1a917a786805342ef71e1379'}]}, 'timestamp': '2025-11-27 10:03:44.169780', '_unique_id': '5787e648ed984229820911f2e2f2cc8c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.171 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.172 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.172 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4f5c8448-f7e6-4a01-82b9-fb44a0f8e6eb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:03:44.172135', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '5b988d34-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.354343447, 'message_signature': '5b76aced2c06fcd0106d5514a3fefa0a1bbdfe82a83b0dd7ff0044baca0cb26b'}]}, 'timestamp': '2025-11-27 10:03:44.172619', '_unique_id': '3597f1532a3d4383ae5afb1684d3cee6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.175 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.176 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9206387e-fc07-413f-a908-7510aa56bbbe', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.175987', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b992442-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '3dbd90ad2baea07aa620c7f16ac6639aafeb84afca3b151b50304d4d43e0c051'}]}, 'timestamp': '2025-11-27 10:03:44.176489', '_unique_id': '55895be024954dc6b11b08f27784025e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.178 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.178 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '81325130-2920-47db-956b-9a6193272eda', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.178939', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b999710-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '54c31b7466ab15e6951e64d9ccf5824fb960df8e9d1daa5860fe412a6128510f'}]}, 'timestamp': '2025-11-27 10:03:44.179409', '_unique_id': 'fc96ba47c53545e4810eb8db283ad145'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.180 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.181 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.181 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '46986725-257e-42ce-86cc-0e7470a8dcaf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.181563', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5b99fd86-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '6e1179e0a762d123888ab49383872c2eb62e8da82f139001baf6152a2f4fcb1b'}]}, 'timestamp': '2025-11-27 10:03:44.182060', '_unique_id': '839ca6c1bccb46e0a81fbef915b0bc46'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.184 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.194 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.194 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c5c67388-2199-4413-9ce1-defb19b84046', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.184191', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5b9bf136-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': 'd0139dea17410e8a60f063d0f1ec01f9333edc21121bce57c926e76798369ba7'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.184191', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5b9c0248-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': '18d50a15de8de4f20bfaead9fbf2723bea604df5afa40152a9881a50ef510fea'}]}, 'timestamp': '2025-11-27 10:03:44.195232', '_unique_id': 'b104588a369443eaaeca726cfd717709'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.196 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.197 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:03:44 localhost nova_compute[284026]: 2025-11-27 10:03:44.204 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.225 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '65793f9e-d7d9-496d-843a-74cc30089bc5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.197630', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba0b6da-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '4f098265412feb8e4cc1a286e7f627f699a8cef4714d656ed90858e565849513'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.197630', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba0cc4c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '5cf7c9d00e54a2a05056fcbef48f8e424f93d62525ed66f8b495747c615c0833'}]}, 'timestamp': '2025-11-27 10:03:44.226703', '_unique_id': '7a5813ace82245a7ad5e8644bf97799e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.227 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.229 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.229 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4cdc33e8-6974-4153-9042-f16767ecca5c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.229228', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5ba148ac-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '628b5e95302ae5746ec847578eab85f0dee05ac650229be5e9a706422f9215e9'}]}, 'timestamp': '2025-11-27 10:03:44.229871', '_unique_id': '14eab41d8e7b440ca954c392b152f5a9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.232 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.232 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.232 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '339a517b-b5ee-44eb-9b75-ef503f5d26ff', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.232220', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba1bcba-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '551f6fd52fb42b01d336e575295eeb5d8eafd9baaddf0144ae537b6b7ccca729'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.232220', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba1d06a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': 'a29950973f4b0a23f6f180b974f4118c73e9228967910f7d038e31e6f2823c40'}]}, 'timestamp': '2025-11-27 10:03:44.233292', '_unique_id': '4c587978f238435499f7c00208ecfa5d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.235 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.235 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '998ff447-dac2-4626-bf73-ab95adade911', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.235758', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5ba24644-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '34a3184db38f9600b03d76b1f180fbf5e8819a63199935673e01e70395b8619d'}]}, 'timestamp': '2025-11-27 10:03:44.236359', '_unique_id': '0e7f12bcd34d43f9816628cf61707b6b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.238 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.238 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b6d8c79e-c9bb-428c-82db-a6e9a885f363', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:03:44.239151', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5ba2cb1e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.326055561, 'message_signature': '5d0b91f86614ae3a4ac726049d68a725cf6374d99645d6826c3b63355d8a422a'}]}, 'timestamp': '2025-11-27 10:03:44.239789', '_unique_id': 'f0dfaa0479df45a0aa87e7812c1ba593'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.241 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '76f26469-13de-4bbb-a7a0-06bad86be993', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.242051', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba33ca2-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '8e2e4fdc7511c3ed654ee5e584b281848581efb9f7ca7bf42cf6bf809a06539d'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.242051', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba352c8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': 'bfdb9dc3e662569c5000b61f85585fdb8ce0379cc0e9fa8339e9ee1a52b3dc5d'}]}, 'timestamp': '2025-11-27 10:03:44.243183', '_unique_id': 'adb0e0b981254e79b575f2299f4452f6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.246 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '5111eb92-e29c-4cc4-85aa-3d25a5ba3c61', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.245484', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba3c320-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': 'ff1cf715363fe0a71693a506bd830955e29fe4538b6c2d0f32c0da7776bd5eeb'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.245484', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba3d784-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': 'a94e83c34eaee04bc3d5a9d3462323b37f3b453287f706137f3f9be59e26cd5b'}]}, 'timestamp': '2025-11-27 10:03:44.246637', '_unique_id': '2fdbfcbbb12f4d61b4e112a9fdf7bbb6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.249 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd1fedfb2-5910-4c39-973e-bcddd31d30ca', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.249211', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba45420-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '7ab2d04fc91308f4766f08b6aa7a651d7214abdb68419645d56488a9cab3215d'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.249211', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba46834-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '834fe0afc8d758798fd27f077d34bc6e3b7d7c7559d04f72d9e9090dcb2d60db'}]}, 'timestamp': '2025-11-27 10:03:44.250312', '_unique_id': '26371f35217f466e90e2730fe9ad2a62'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.252 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.252 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e2c78918-854b-4f87-8915-90e608431be9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.253149', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba4ed72-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': '3a010cf67ed538a9c0424723789abf1223aae359972188d9efaad11ee9d9f771'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.253149', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba50212-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.372253913, 'message_signature': '88fac81703c3ee1b5326c19b638feb499857b7771c0844cec494bda6541138e8'}]}, 'timestamp': '2025-11-27 10:03:44.254252', '_unique_id': '0fd719cfe9714ac18c1c921b234d3e20'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8ac90838-b634-4477-9434-2ae21a46977d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.256759', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba577c4-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '75c6daf0f858ec2ce4c100465a46e50b04488f77955724c78eb4e74c969cee53'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.256759', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba58c78-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': '5c58fef183a72751f79c25793f9f364cc8facc2b7d06c7d946018a892276efec'}]}, 'timestamp': '2025-11-27 10:03:44.257754', '_unique_id': '70d5397fc5884a609a16401011e31daa'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'fd5125aa-48ae-4313-8a3c-4cdef34189c3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:03:44.259611', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5ba5e254-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': 'd3b35b9b47643406d04c94767b1920363bd3d550217ba0f1fcd46d194c04c116'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:03:44.259611', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5ba5ebd2-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12054.385700206, 'message_signature': 'e7f36cf95404d4d30f863383df27010028f68bcab27d12f27857e04ceaef076c'}]}, 'timestamp': '2025-11-27 10:03:44.260117', '_unique_id': 'f0bb91744ae34779bde35d540651c874'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:03:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:03:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:03:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:03:44 Nov 27 05:03:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:03:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:03:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'manila_data', 'vms', 'manila_metadata', 'images', '.mgr', 'volumes'] Nov 27 05:03:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:03:44 localhost nova_compute[284026]: 2025-11-27 10:03:44.446 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:03:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0048184716987716356 of space, bias 1.0, pg target 0.9636943397543272 quantized to 32 (current 32) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8570103846780196 quantized to 32 (current 32) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:03:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002166503815373162 quantized to 16 (current 16) Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:03:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:03:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v126: 177 pgs: 177 active+clean; 192 MiB data, 788 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 12 KiB/s wr, 73 op/s Nov 27 05:03:45 localhost nova_compute[284026]: 2025-11-27 10:03:45.402 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:45.420 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:44Z, description=, device_id=74490902-5f1a-410c-92b8-656d5f249d38, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=8b9090e7-88ef-437a-80c7-02cdbbf68e8e, ip_allocation=immediate, mac_address=fa:16:3e:d9:81:99, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=586, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:03:45Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:03:45 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 7 addresses Nov 27 05:03:45 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:03:45 localhost podman[311666]: 2025-11-27 10:03:45.659715406 +0000 UTC m=+0.072646509 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:03:45 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:03:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:45.959 265123 INFO neutron.agent.dhcp.agent [None req-27c1032c-6d75-474d-a6c7-16176794e4c4 - - - - - -] DHCP configuration for ports {'8b9090e7-88ef-437a-80c7-02cdbbf68e8e'} is completed#033[00m Nov 27 05:03:46 localhost nova_compute[284026]: 2025-11-27 10:03:46.011 284030 DEBUG nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] pre_live_migration data is LibvirtLiveMigrateData(bdms=,block_migration=False,disk_available_mb=13312,disk_over_commit=,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp110wzs4o',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='ea7c65ed-7232-403d-84e6-2ead60aade7c',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=,old_vol_attachment_ids=,serial_listen_addr=None,serial_listen_ports=,src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=,target_connect_addr=,vifs=[VIFMigrateData],wait_for_vif_plugged=) pre_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:8604#033[00m Nov 27 05:03:46 localhost nova_compute[284026]: 2025-11-27 10:03:46.202 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquiring lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:03:46 localhost nova_compute[284026]: 2025-11-27 10:03:46.203 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquired lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:03:46 localhost nova_compute[284026]: 2025-11-27 10:03:46.203 284030 DEBUG nova.network.neutron [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Building network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2010#033[00m Nov 27 05:03:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:46.282 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:45Z, description=, device_id=ef2bcfd0-753b-4017-b94e-fb3e2a8510cb, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=baa1573b-0b8a-487d-ae29-28005ea4f98d, ip_allocation=immediate, mac_address=fa:16:3e:2b:b6:78, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:03:33Z, description=, dns_domain=, id=75d79a30-b416-43bb-83b3-cd22cce0e128, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-LiveMigrationTest-1063024354-network, port_security_enabled=True, project_id=2f8a8262d8074ffd876910fbc25620fe, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=51747, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=542, status=ACTIVE, subnets=['265a20a8-eadc-4e3a-a919-237de22bfc6c'], tags=[], tenant_id=2f8a8262d8074ffd876910fbc25620fe, updated_at=2025-11-27T10:03:38Z, vlan_transparent=None, network_id=75d79a30-b416-43bb-83b3-cd22cce0e128, port_security_enabled=False, project_id=2f8a8262d8074ffd876910fbc25620fe, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=593, status=DOWN, tags=[], tenant_id=2f8a8262d8074ffd876910fbc25620fe, updated_at=2025-11-27T10:03:45Z on network 75d79a30-b416-43bb-83b3-cd22cce0e128#033[00m Nov 27 05:03:46 localhost dnsmasq[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/addn_hosts - 1 addresses Nov 27 05:03:46 localhost podman[311705]: 2025-11-27 10:03:46.647620736 +0000 UTC m=+0.067519960 container kill 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:03:46 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/host Nov 27 05:03:46 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/opts Nov 27 05:03:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:46.936 265123 INFO neutron.agent.dhcp.agent [None req-917e7f44-9dcc-4d0c-a41c-448472af9f64 - - - - - -] DHCP configuration for ports {'baa1573b-0b8a-487d-ae29-28005ea4f98d'} is completed#033[00m Nov 27 05:03:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v127: 177 pgs: 177 active+clean; 238 MiB data, 856 MiB used, 41 GiB / 42 GiB avail; 3.6 MiB/s rd, 1.8 MiB/s wr, 117 op/s Nov 27 05:03:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.390 284030 DEBUG nova.network.neutron [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Updating instance_info_cache with network_info: [{"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.425 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.440 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Releasing lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.442 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] migrate_data in pre_live_migration: LibvirtLiveMigrateData(bdms=,block_migration=False,disk_available_mb=13312,disk_over_commit=,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp110wzs4o',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='ea7c65ed-7232-403d-84e6-2ead60aade7c',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=,old_vol_attachment_ids={},serial_listen_addr=None,serial_listen_ports=,src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=,target_connect_addr=,vifs=[VIFMigrateData],wait_for_vif_plugged=) pre_live_migration /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10827#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.442 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Creating instance directory: /var/lib/nova/instances/ea7c65ed-7232-403d-84e6-2ead60aade7c pre_live_migration /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10840#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.443 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Ensure instance console log exists: /var/lib/nova/instances/ea7c65ed-7232-403d-84e6-2ead60aade7c/console.log _ensure_console_log_for_instance /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4609#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.443 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Plugging VIFs using destination host port bindings before live migration. _pre_live_migration_plug_vifs /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10794#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.446 284030 DEBUG nova.virt.libvirt.vif [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='True',created_at=2025-11-27T10:03:11Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description=None,display_name='tempest-LiveAutoBlockMigrationV225Test-server-1085268656',ec2_ids=,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537444.localdomain',hostname='tempest-liveautoblockmigrationv225test-server-1085268656',id=6,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T10:03:37Z,launched_on='np0005537444.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=,new_flavor=None,node='np0005537444.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=InstancePCIRequests,power_state=1,progress=0,project_id='0780cc2b55f74fbd8c97a1f5f39cd8cc',ramdisk_id='',reservation_id='r-f7ny7m85',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='q35',image_hw_pointer_model='usbtablet',image_hw_rng_model='virtio',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',owner_project_name='tempest-LiveAutoBlockMigrationV225Test-1706287954',owner_user_name='tempest-LiveAutoBlockMigrationV225Test-1706287954-project-member'},tags=,task_state='migrating',terminated_at=None,trusted_certs=,updated_at=2025-11-27T10:03:37Z,user_data=None,user_id='730775ba58804b5b8977106a8a8aa97c',uuid=ea7c65ed-7232-403d-84e6-2ead60aade7c,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system"}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.447 284030 DEBUG nova.network.os_vif_util [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Converting VIF {"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system"}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.448 284030 DEBUG nova.network.os_vif_util [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.449 284030 DEBUG os_vif [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Plugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.451 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.452 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.452 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.457 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.457 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap30bcf056-09, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.458 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tap30bcf056-09, col_values=(('external_ids', {'iface-id': '30bcf056-09f4-4ccc-94c7-ca6d81a754e9', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:83:7f:f2', 'vm-uuid': 'ea7c65ed-7232-403d-84e6-2ead60aade7c'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.463 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.467 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.469 284030 INFO os_vif [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Successfully plugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09')#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.470 284030 DEBUG nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] No dst_numa_info in migrate_data, no cores to power up in pre_live_migration. pre_live_migration /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10954#033[00m Nov 27 05:03:47 localhost nova_compute[284026]: 2025-11-27 10:03:47.470 284030 DEBUG nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] driver pre_live_migration data is LibvirtLiveMigrateData(bdms=[],block_migration=False,disk_available_mb=13312,disk_over_commit=,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp110wzs4o',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='ea7c65ed-7232-403d-84e6-2ead60aade7c',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=,old_vol_attachment_ids={},serial_listen_addr=None,serial_listen_ports=[],src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=[],target_connect_addr=None,vifs=[VIFMigrateData],wait_for_vif_plugged=) pre_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:8668#033[00m Nov 27 05:03:48 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:48.077 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '10'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v128: 177 pgs: 177 active+clean; 238 MiB data, 856 MiB used, 41 GiB / 42 GiB avail; 3.6 MiB/s rd, 1.8 MiB/s wr, 107 op/s Nov 27 05:03:50 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:50.230 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:03:45Z, description=, device_id=ef2bcfd0-753b-4017-b94e-fb3e2a8510cb, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=baa1573b-0b8a-487d-ae29-28005ea4f98d, ip_allocation=immediate, mac_address=fa:16:3e:2b:b6:78, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:03:33Z, description=, dns_domain=, id=75d79a30-b416-43bb-83b3-cd22cce0e128, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-LiveMigrationTest-1063024354-network, port_security_enabled=True, project_id=2f8a8262d8074ffd876910fbc25620fe, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=51747, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=542, status=ACTIVE, subnets=['265a20a8-eadc-4e3a-a919-237de22bfc6c'], tags=[], tenant_id=2f8a8262d8074ffd876910fbc25620fe, updated_at=2025-11-27T10:03:38Z, vlan_transparent=None, network_id=75d79a30-b416-43bb-83b3-cd22cce0e128, port_security_enabled=False, project_id=2f8a8262d8074ffd876910fbc25620fe, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=593, status=DOWN, tags=[], tenant_id=2f8a8262d8074ffd876910fbc25620fe, updated_at=2025-11-27T10:03:45Z on network 75d79a30-b416-43bb-83b3-cd22cce0e128#033[00m Nov 27 05:03:50 localhost nova_compute[284026]: 2025-11-27 10:03:50.404 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:50 localhost dnsmasq[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/addn_hosts - 1 addresses Nov 27 05:03:50 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/host Nov 27 05:03:50 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/opts Nov 27 05:03:50 localhost podman[311745]: 2025-11-27 10:03:50.556598411 +0000 UTC m=+0.056151442 container kill 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:03:50 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:03:50.826 265123 INFO neutron.agent.dhcp.agent [None req-63f69081-286a-4e9e-a90f-cbee4e7d3ffc - - - - - -] DHCP configuration for ports {'baa1573b-0b8a-487d-ae29-28005ea4f98d'} is completed#033[00m Nov 27 05:03:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v129: 177 pgs: 177 active+clean; 264 MiB data, 926 MiB used, 41 GiB / 42 GiB avail; 5.8 MiB/s rd, 3.8 MiB/s wr, 217 op/s Nov 27 05:03:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:52 localhost nova_compute[284026]: 2025-11-27 10:03:52.427 284030 DEBUG nova.network.neutron [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Port 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 updated with migration profile {'migrating_to': 'np0005537446.localdomain'} successfully _setup_migration_port_profile /usr/lib/python3.9/site-packages/nova/network/neutron.py:354#033[00m Nov 27 05:03:52 localhost nova_compute[284026]: 2025-11-27 10:03:52.429 284030 DEBUG nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] pre_live_migration result data is LibvirtLiveMigrateData(bdms=[],block_migration=False,disk_available_mb=13312,disk_over_commit=,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp110wzs4o',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='ea7c65ed-7232-403d-84e6-2ead60aade7c',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=,old_vol_attachment_ids={},serial_listen_addr=None,serial_listen_ports=[],src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=[],target_connect_addr=None,vifs=[VIFMigrateData],wait_for_vif_plugged=True) pre_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:8723#033[00m Nov 27 05:03:52 localhost nova_compute[284026]: 2025-11-27 10:03:52.462 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:52 localhost sshd[311765]: main: sshd: ssh-rsa algorithm is disabled Nov 27 05:03:52 localhost systemd-logind[761]: New session 74 of user nova. Nov 27 05:03:52 localhost systemd[1]: Created slice User Slice of UID 42436. Nov 27 05:03:52 localhost systemd[1]: Starting User Runtime Directory /run/user/42436... Nov 27 05:03:52 localhost systemd[1]: Finished User Runtime Directory /run/user/42436. Nov 27 05:03:52 localhost systemd[1]: Starting User Manager for UID 42436... Nov 27 05:03:53 localhost systemd[311769]: Queued start job for default target Main User Target. Nov 27 05:03:53 localhost systemd[311769]: Created slice User Application Slice. Nov 27 05:03:53 localhost systemd[311769]: Started Mark boot as successful after the user session has run 2 minutes. Nov 27 05:03:53 localhost systemd[311769]: Started Daily Cleanup of User's Temporary Directories. Nov 27 05:03:53 localhost systemd[311769]: Reached target Paths. Nov 27 05:03:53 localhost systemd[311769]: Reached target Timers. Nov 27 05:03:53 localhost systemd[311769]: Starting D-Bus User Message Bus Socket... Nov 27 05:03:53 localhost systemd[311769]: Starting Create User's Volatile Files and Directories... Nov 27 05:03:53 localhost systemd[311769]: Listening on D-Bus User Message Bus Socket. Nov 27 05:03:53 localhost systemd[311769]: Reached target Sockets. Nov 27 05:03:53 localhost systemd[311769]: Finished Create User's Volatile Files and Directories. Nov 27 05:03:53 localhost systemd[311769]: Reached target Basic System. Nov 27 05:03:53 localhost systemd[311769]: Reached target Main User Target. Nov 27 05:03:53 localhost systemd[311769]: Startup finished in 173ms. Nov 27 05:03:53 localhost systemd[1]: Started User Manager for UID 42436. Nov 27 05:03:53 localhost systemd[1]: Started Session 74 of User nova. Nov 27 05:03:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v130: 177 pgs: 177 active+clean; 264 MiB data, 926 MiB used, 41 GiB / 42 GiB avail; 3.9 MiB/s rd, 3.8 MiB/s wr, 153 op/s Nov 27 05:03:53 localhost systemd[1]: Started libvirt secret daemon. Nov 27 05:03:53 localhost kernel: device tap30bcf056-09 entered promiscuous mode Nov 27 05:03:53 localhost NetworkManager[5971]: [1764237833.2949] manager: (tap30bcf056-09): new Tun device (/org/freedesktop/NetworkManager/Devices/20) Nov 27 05:03:53 localhost systemd-udevd[311818]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.292 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00072|binding|INFO|Claiming lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for this additional chassis. Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00073|binding|INFO|30bcf056-09f4-4ccc-94c7-ca6d81a754e9: Claiming fa:16:3e:83:7f:f2 10.100.0.7 Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00074|binding|INFO|Claiming lport 86dea736-4b08-40ac-86cd-1432d75dbf91 for this additional chassis. Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00075|binding|INFO|86dea736-4b08-40ac-86cd-1432d75dbf91: Claiming fa:16:3e:ed:d8:4f 19.80.0.59 Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.300 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost NetworkManager[5971]: [1764237833.3170] device (tap30bcf056-09): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 05:03:53 localhost NetworkManager[5971]: [1764237833.3177] device (tap30bcf056-09): state change: unavailable -> disconnected (reason 'none', sys-iface-state: 'external') Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.336 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00076|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:03:53 localhost systemd-machined[86026]: New machine qemu-3-instance-00000006. Nov 27 05:03:53 localhost systemd[1]: Started Virtual Machine qemu-3-instance-00000006. Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.368 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost ovn_controller[156436]: 2025-11-27T10:03:53Z|00077|binding|INFO|Setting lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 ovn-installed in OVS Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.384 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.685 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Started> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.686 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] VM Started (Lifecycle Event)#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.718 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.915 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.937 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Resumed> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.937 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] VM Resumed (Lifecycle Event)#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.982 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:03:53 localhost nova_compute[284026]: 2025-11-27 10:03:53.988 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Synchronizing instance power state after lifecycle event "Resumed"; current vm_state: active, current task_state: migrating, current DB power_state: 1, VM power_state: 1 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 05:03:54 localhost nova_compute[284026]: 2025-11-27 10:03:54.027 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] During the sync_power process the instance has moved from host np0005537444.localdomain to host np0005537446.localdomain#033[00m Nov 27 05:03:54 localhost systemd[1]: session-74.scope: Deactivated successfully. Nov 27 05:03:54 localhost systemd-logind[761]: Session 74 logged out. Waiting for processes to exit. Nov 27 05:03:54 localhost systemd-logind[761]: Removed session 74. Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00078|binding|INFO|Claiming lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for this chassis. Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00079|binding|INFO|30bcf056-09f4-4ccc-94c7-ca6d81a754e9: Claiming fa:16:3e:83:7f:f2 10.100.0.7 Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00080|binding|INFO|Claiming lport 86dea736-4b08-40ac-86cd-1432d75dbf91 for this chassis. Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00081|binding|INFO|86dea736-4b08-40ac-86cd-1432d75dbf91: Claiming fa:16:3e:ed:d8:4f 19.80.0.59 Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00082|binding|INFO|Setting lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 up in Southbound Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00083|binding|INFO|Setting lport 86dea736-4b08-40ac-86cd-1432d75dbf91 up in Southbound Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.034 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:ed:d8:4f 19.80.0.59'], port_security=['fa:16:3e:ed:d8:4f 19.80.0.59'], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': ''}, parent_port=['30bcf056-09f4-4ccc-94c7-ca6d81a754e9'], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-subport-447612827', 'neutron:cidrs': '19.80.0.59/24', 'neutron:device_id': '', 'neutron:device_owner': 'trunk:subport', 'neutron:mtu': '', 'neutron:network_name': 'neutron-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-subport-447612827', 'neutron:project_id': '0780cc2b55f74fbd8c97a1f5f39cd8cc', 'neutron:revision_number': '3', 'neutron:security_group_ids': 'd5a4dbad-2a9b-45ce-81ca-0430479e3c01', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[42], additional_encap=[], encap=[], mirror_rules=[], datapath=174951dd-7028-48c6-9b6e-de8f9b584ff5, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=86dea736-4b08-40ac-86cd-1432d75dbf91) old=Port_Binding(up=[False], additional_chassis=[], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.037 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:83:7f:f2 10.100.0.7'], port_security=['fa:16:3e:83:7f:f2 10.100.0.7'], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-parent-1295767349', 'neutron:cidrs': '10.100.0.7/28', 'neutron:device_id': 'ea7c65ed-7232-403d-84e6-2ead60aade7c', 'neutron:device_owner': 'compute:nova', 'neutron:mtu': '', 'neutron:network_name': 'neutron-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-parent-1295767349', 'neutron:project_id': '0780cc2b55f74fbd8c97a1f5f39cd8cc', 'neutron:revision_number': '9', 'neutron:security_group_ids': 'd5a4dbad-2a9b-45ce-81ca-0430479e3c01', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537444.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=1cc89ba8-681c-4c44-a43f-cbe7c9b9d281, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=30bcf056-09f4-4ccc-94c7-ca6d81a754e9) old=Port_Binding(up=[False], additional_chassis=[], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.039 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 86dea736-4b08-40ac-86cd-1432d75dbf91 in datapath 89c93072-e57c-4d59-b8c2-3bb1a82cf80c bound to our chassis#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.043 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network 89c93072-e57c-4d59-b8c2-3bb1a82cf80c#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.055 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[50bee2fa-bf7b-4ab1-8267-d9c491d457a4]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.056 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Creating VETH tap89c93072-e1 in ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c namespace provision_datapath /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:665#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.058 162196 DEBUG neutron.privileged.agent.linux.ip_lib [-] Interface tap89c93072-e0 not found in namespace None get_link_id /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:204#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.058 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[5f009f03-d2c2-4038-81ed-12f7c6f914a5]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.060 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a8eb0c4b-fdca-4dc2-a865-a7bf1787031e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.079 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[f5508192-e043-4f43-a140-2d8d464be949]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.093 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[18cc4b6e-26b0-4647-8a0f-186d91f25912]: (4, ('net.ipv4.conf.all.promote_secondaries = 1\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v131: 177 pgs: 177 active+clean; 264 MiB data, 926 MiB used, 41 GiB / 42 GiB avail; 3.9 MiB/s rd, 3.8 MiB/s wr, 153 op/s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.123 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[ea41e1ac-cfcc-44af-9677-d9fa1308f448]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost NetworkManager[5971]: [1764237835.1328] manager: (tap89c93072-e0): new Veth device (/org/freedesktop/NetworkManager/Devices/21) Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.131 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[dc6dfa12-a74b-4330-83f8-2297fa711ec7]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost neutron_sriov_agent[258162]: 2025-11-27 10:03:55.138 2 WARNING neutron.plugins.ml2.drivers.mech_sriov.agent.sriov_nic_agent [req-24fee2f0-ef00-42d2-985e-28fd86d881d4 req-2abcbe97-7c6a-41aa-8efa-f66290dd7752 f710e6626a0e4b49b815f0fd6bdc2fb3 ece81c3214184c54956c375af05c2f4f - - default default] This port is not SRIOV, skip binding for port 30bcf056-09f4-4ccc-94c7-ca6d81a754e9.#033[00m Nov 27 05:03:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.168 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[43228a9a-4524-42e0-9197-71deb1013978]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.174 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[395928ef-740e-4e59-b2d7-40b2bcf51fa0]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap89c93072-e1: link becomes ready Nov 27 05:03:55 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap89c93072-e0: link becomes ready Nov 27 05:03:55 localhost NetworkManager[5971]: [1764237835.2031] device (tap89c93072-e0): carrier: link connected Nov 27 05:03:55 localhost systemd[1]: tmp-crun.0xTOWr.mount: Deactivated successfully. Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.212 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[0d637655-2c63-4e6a-b134-39f3bf68b1a5]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost podman[311877]: 2025-11-27 10:03:55.219673871 +0000 UTC m=+0.111617305 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2) Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.230 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[209c4ce6-022c-45e8-9f23-3815478a627d]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap89c93072-e1'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:ed:af:1e'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 22], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206531, 'reachable_time': 25961, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 311922, 'error': None, 'target': 'ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost podman[311877]: 2025-11-27 10:03:55.23698628 +0000 UTC m=+0.128929674 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2) Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.247 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[bc3782bf-dcef-45bd-a9a0-50dd6f160a57]: (4, ({'family': 10, 'prefixlen': 64, 'flags': 192, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:feed:af1e'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 1206531, 'tstamp': 1206531}], ['IFA_FLAGS', 192]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 311928, 'error': None, 'target': 'ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'},)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.267 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e36483a5-1d7b-4dd2-885a-9dbd8733d303]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap89c93072-e1'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:ed:af:1e'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 22], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206531, 'reachable_time': 25961, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 0, 'sequence_number': 255, 'pid': 311931, 'error': None, 'target': 'ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.300 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[74a159da-28b4-4460-895a-0a52e612010b]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost neutron_sriov_agent[258162]: 2025-11-27 10:03:55.332 2 INFO neutron.agent.securitygroups_rpc [None req-7a2aea18-16b1-4fbb-b52a-8c47f6fe9ac8 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Security group member updated ['6d11bb1f-93da-41d7-8007-c66992d17e28']#033[00m Nov 27 05:03:55 localhost podman[311908]: 2025-11-27 10:03:55.333237637 +0000 UTC m=+0.152397579 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ovn_metadata_agent, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.358 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3a0764fe-3145-4c06-b0de-8fd19db625f8]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.360 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap89c93072-e0, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.360 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.361 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap89c93072-e0, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.363 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:55 localhost kernel: device tap89c93072-e0 entered promiscuous mode Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.368 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tap89c93072-e0, col_values=(('external_ids', {'iface-id': 'cdeccc19-a6f6-4ac9-b28c-bcd5b506ddbe'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.369 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:55 localhost ovn_controller[156436]: 2025-11-27T10:03:55Z|00084|binding|INFO|Releasing lport cdeccc19-a6f6-4ac9-b28c-bcd5b506ddbe from this chassis (sb_readonly=0) Nov 27 05:03:55 localhost podman[311908]: 2025-11-27 10:03:55.37395875 +0000 UTC m=+0.193118652 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.378 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.380 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/89c93072-e57c-4d59-b8c2-3bb1a82cf80c.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/89c93072-e57c-4d59-b8c2-3bb1a82cf80c.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.381 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d0b8069e-262b-4c58-b782-45ce67353767]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.382 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: global Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-89c93072-e57c-4d59-b8c2-3bb1a82cf80c Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: user root Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: group root Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/89c93072-e57c-4d59-b8c2-3bb1a82cf80c.pid.haproxy Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: daemon Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: defaults Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: log global Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: mode http Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: option httplog Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: listen listener Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID 89c93072-e57c-4d59-b8c2-3bb1a82cf80c Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.382 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'env', 'PROCESS_TAG=haproxy-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/89c93072-e57c-4d59-b8c2-3bb1a82cf80c.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 05:03:55 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.408 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.433 284030 INFO nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Post operation of migration started#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.589 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquiring lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.589 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquired lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:03:55 localhost nova_compute[284026]: 2025-11-27 10:03:55.590 284030 DEBUG nova.network.neutron [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Building network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2010#033[00m Nov 27 05:03:55 localhost openstack_network_exporter[244641]: ERROR 10:03:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:03:55 localhost openstack_network_exporter[244641]: ERROR 10:03:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:03:55 localhost openstack_network_exporter[244641]: ERROR 10:03:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:03:55 localhost openstack_network_exporter[244641]: ERROR 10:03:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:03:55 localhost openstack_network_exporter[244641]: Nov 27 05:03:55 localhost openstack_network_exporter[244641]: ERROR 10:03:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:03:55 localhost openstack_network_exporter[244641]: Nov 27 05:03:55 localhost podman[311972]: Nov 27 05:03:55 localhost podman[311972]: 2025-11-27 10:03:55.81068533 +0000 UTC m=+0.096219368 container create a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:03:55 localhost systemd[1]: Started libpod-conmon-a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5.scope. Nov 27 05:03:55 localhost podman[311972]: 2025-11-27 10:03:55.76009643 +0000 UTC m=+0.045630508 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 05:03:55 localhost systemd[1]: Started libcrun container. Nov 27 05:03:55 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c888ae8a8f2d5df990e7266ba4dbb42cc2b9a0b65d0ce1dc20c878b8f1e7170b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:03:55 localhost podman[311972]: 2025-11-27 10:03:55.879092533 +0000 UTC m=+0.164626581 container init a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:03:55 localhost podman[311972]: 2025-11-27 10:03:55.888503598 +0000 UTC m=+0.174037636 container start a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:03:55 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [NOTICE] (311990) : New worker (312000) forked Nov 27 05:03:55 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [NOTICE] (311990) : Loading success. Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.954 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 in datapath f966841f-0fe6-48fb-bc34-0c4a63a8d9fc bound to our chassis#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.958 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network f966841f-0fe6-48fb-bc34-0c4a63a8d9fc#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.971 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[1327964d-ecbd-4457-9e2d-5d7d8b1f86bc]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.975 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Creating VETH tapf966841f-01 in ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc namespace provision_datapath /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:665#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.978 162196 DEBUG neutron.privileged.agent.linux.ip_lib [-] Interface tapf966841f-00 not found in namespace None get_link_id /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:204#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.979 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a63d12ca-6fb2-40c9-9532-414e3b8e88e9]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.980 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c4ef738e-b3bb-465d-bcff-60e8c79bbcac]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:55.989 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[8ed7fbad-b556-480c-bf98-548a57fb05a0]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.007 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8780daa1-703c-4eba-a252-ca8209a74c1f]: (4, ('net.ipv4.conf.all.promote_secondaries = 1\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.044 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[87e61b41-1522-4617-bd0c-cc96105341a0]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost NetworkManager[5971]: [1764237836.0569] manager: (tapf966841f-00): new Veth device (/org/freedesktop/NetworkManager/Devices/22) Nov 27 05:03:56 localhost systemd-udevd[311907]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.054 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[adb13bda-2f66-4ac2-bee7-13b740d1d75c]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.104 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[38ac13e0-32b6-4c56-8e39-4b51d3d03794]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.109 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[a135773b-44a1-4591-85b4-f32371d2798c]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost NetworkManager[5971]: [1764237836.1373] device (tapf966841f-00): carrier: link connected Nov 27 05:03:56 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tapf966841f-00: link becomes ready Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.141 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[d20b7268-1909-4658-a437-e717cfaf72b7]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.162 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3f001dad-b250-4111-affb-167cc71652a0]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tapf966841f-01'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:b0:e0:b4'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 23], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206625, 'reachable_time': 19584, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 312047, 'error': None, 'target': 'ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.180 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[72087f80-8ac5-4b2b-a661-74d0a167257c]: (4, ({'family': 10, 'prefixlen': 64, 'flags': 192, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:feb0:e0b4'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 1206625, 'tstamp': 1206625}], ['IFA_FLAGS', 192]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 312049, 'error': None, 'target': 'ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'},)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.186 284030 DEBUG nova.network.neutron [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Updating instance_info_cache with network_info: [{"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.198 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[fae4716d-778f-408d-b599-9dbf7df43597]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tapf966841f-01'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:b0:e0:b4'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 23], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206625, 'reachable_time': 19584, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 0, 'sequence_number': 255, 'pid': 312050, 'error': None, 'target': 'ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.229 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c312b0dc-73f0-4450-96ff-564497992d2e]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.281 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Releasing lock "refresh_cache-ea7c65ed-7232-403d-84e6-2ead60aade7c" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.283 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[023f1893-3cc0-47d3-a731-856e5937ead3]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.285 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tapf966841f-00, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.286 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.287 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tapf966841f-00, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.290 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:56 localhost kernel: device tapf966841f-00 entered promiscuous mode Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.293 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.295 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tapf966841f-00, col_values=(('external_ids', {'iface-id': 'b5e8fcfd-584f-4244-9345-c50b7c0b100b'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.297 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:56 localhost ovn_controller[156436]: 2025-11-27T10:03:56Z|00085|binding|INFO|Releasing lport b5e8fcfd-584f-4244-9345-c50b7c0b100b from this chassis (sb_readonly=0) Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.302 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.allocate_pci_devices_for_instance" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.302 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.allocate_pci_devices_for_instance" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.303 284030 DEBUG oslo_concurrency.lockutils [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.allocate_pci_devices_for_instance" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.306 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.308 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/f966841f-0fe6-48fb-bc34-0c4a63a8d9fc.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/f966841f-0fe6-48fb-bc34-0c4a63a8d9fc.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.309 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[052508ea-89a2-4895-80e0-ff3147a9250a]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.310 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: global Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: user root Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: group root Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/f966841f-0fe6-48fb-bc34-0c4a63a8d9fc.pid.haproxy Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: daemon Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: defaults Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: log global Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: mode http Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: option httplog Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: listen listener Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID f966841f-0fe6-48fb-bc34-0c4a63a8d9fc Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 05:03:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:03:56.312 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'env', 'PROCESS_TAG=haproxy-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/f966841f-0fe6-48fb-bc34-0c4a63a8d9fc.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 05:03:56 localhost nova_compute[284026]: 2025-11-27 10:03:56.316 284030 INFO nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Sending announce-self command to QEMU monitor. Attempt 1 of 3#033[00m Nov 27 05:03:56 localhost journal[205316]: Domain id=3 name='instance-00000006' uuid=ea7c65ed-7232-403d-84e6-2ead60aade7c is tainted: custom-monitor Nov 27 05:03:56 localhost podman[312099]: Nov 27 05:03:56 localhost podman[312099]: 2025-11-27 10:03:56.743630871 +0000 UTC m=+0.084315665 container create 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:03:56 localhost systemd[1]: Started libpod-conmon-3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed.scope. Nov 27 05:03:56 localhost podman[312099]: 2025-11-27 10:03:56.69783927 +0000 UTC m=+0.038524124 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 05:03:56 localhost systemd[1]: Started libcrun container. Nov 27 05:03:56 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/fe2810b887e4c6ff976b50fdb48de78b18e4541fb2a46d627acc143cdfdba828/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:03:56 localhost podman[312099]: 2025-11-27 10:03:56.838209833 +0000 UTC m=+0.178894647 container init 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:03:56 localhost podman[312099]: 2025-11-27 10:03:56.84438577 +0000 UTC m=+0.185070594 container start 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, tcib_managed=true, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:03:56 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [NOTICE] (312132) : New worker (312134) forked Nov 27 05:03:56 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [NOTICE] (312132) : Loading success. Nov 27 05:03:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:03:56 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:03:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:03:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:03:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:03:56 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 4fdadc1d-83e8-41f9-a932-20e92be2e8b4 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:03:56 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 4fdadc1d-83e8-41f9-a932-20e92be2e8b4 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:03:56 localhost ceph-mgr[290377]: [progress INFO root] Completed event 4fdadc1d-83e8-41f9-a932-20e92be2e8b4 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:03:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:03:56 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:03:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v132: 177 pgs: 177 active+clean; 271 MiB data, 932 MiB used, 41 GiB / 42 GiB avail; 3.9 MiB/s rd, 3.9 MiB/s wr, 172 op/s Nov 27 05:03:57 localhost nova_compute[284026]: 2025-11-27 10:03:57.325 284030 INFO nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Sending announce-self command to QEMU monitor. Attempt 2 of 3#033[00m Nov 27 05:03:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:03:57 localhost nova_compute[284026]: 2025-11-27 10:03:57.464 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:03:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:03:57 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:03:58 localhost nova_compute[284026]: 2025-11-27 10:03:58.333 284030 INFO nova.virt.libvirt.driver [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Sending announce-self command to QEMU monitor. Attempt 3 of 3#033[00m Nov 27 05:03:58 localhost nova_compute[284026]: 2025-11-27 10:03:58.337 284030 DEBUG nova.compute.manager [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:03:58 localhost nova_compute[284026]: 2025-11-27 10:03:58.375 284030 DEBUG nova.objects.instance [None req-24fee2f0-ef00-42d2-985e-28fd86d881d4 1a4e0643e06b46c185a33db37038e0e9 e2b488afb1b7473eaa1d2801328c8b98 - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Trying to apply a migration context that does not seem to be set for this instance apply_migration_context /usr/lib/python3.9/site-packages/nova/objects/instance.py:1032#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.071 284030 DEBUG nova.compute.manager [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.072 284030 DEBUG oslo_concurrency.lockutils [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.073 284030 DEBUG oslo_concurrency.lockutils [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.073 284030 DEBUG oslo_concurrency.lockutils [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.074 284030 DEBUG nova.compute.manager [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] No waiting events found dispatching network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:03:59 localhost nova_compute[284026]: 2025-11-27 10:03:59.074 284030 WARNING nova.compute.manager [req-9f579647-da64-4161-a218-a7d8e5a7b2ad req-9b9d5535-81be-40f4-999b-11ad4c198f70 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received unexpected event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for instance with vm_state active and task_state None.#033[00m Nov 27 05:03:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v133: 177 pgs: 177 active+clean; 271 MiB data, 932 MiB used, 41 GiB / 42 GiB avail; 2.2 MiB/s rd, 2.1 MiB/s wr, 128 op/s Nov 27 05:03:59 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:03:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:04:00 localhost nova_compute[284026]: 2025-11-27 10:04:00.426 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:04:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v134: 177 pgs: 177 active+clean; 304 MiB data, 1008 MiB used, 41 GiB / 42 GiB avail; 2.6 MiB/s rd, 4.3 MiB/s wr, 198 op/s Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.824 284030 DEBUG nova.compute.manager [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.825 284030 DEBUG oslo_concurrency.lockutils [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.826 284030 DEBUG oslo_concurrency.lockutils [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.826 284030 DEBUG oslo_concurrency.lockutils [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.826 284030 DEBUG nova.compute.manager [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] No waiting events found dispatching network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:01 localhost nova_compute[284026]: 2025-11-27 10:04:01.827 284030 WARNING nova.compute.manager [req-dde73495-d030-4082-9635-de341d252693 req-36b209e9-ef29-4bf1-bcf1-e0ee5bf88c15 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received unexpected event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for instance with vm_state active and task_state None.#033[00m Nov 27 05:04:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e91 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:02 localhost nova_compute[284026]: 2025-11-27 10:04:02.466 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost podman[312178]: 2025-11-27 10:04:03.000126232 +0000 UTC m=+0.051801484 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:04:03 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 6 addresses Nov 27 05:04:03 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:03 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00086|binding|INFO|Releasing lport cdeccc19-a6f6-4ac9-b28c-bcd5b506ddbe from this chassis (sb_readonly=0) Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00087|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00088|binding|INFO|Releasing lport b5e8fcfd-584f-4244-9345-c50b7c0b100b from this chassis (sb_readonly=0) Nov 27 05:04:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v135: 177 pgs: 177 active+clean; 304 MiB data, 1008 MiB used, 41 GiB / 42 GiB avail; 456 KiB/s rd, 2.3 MiB/s wr, 88 op/s Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.152 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:03.258 265123 INFO neutron.agent.linux.ip_lib [None req-60d03956-9f02-497e-bb8d-ce4859b68514 - - - - - -] Device tap7d36709c-2f cannot be used as it has no MAC address#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.280 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost kernel: device tap7d36709c-2f entered promiscuous mode Nov 27 05:04:03 localhost NetworkManager[5971]: [1764237843.2912] manager: (tap7d36709c-2f): new Generic device (/org/freedesktop/NetworkManager/Devices/23) Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00089|binding|INFO|Claiming lport 7d36709c-2fe6-487c-9486-157a966792a0 for this chassis. Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00090|binding|INFO|7d36709c-2fe6-487c-9486-157a966792a0: Claiming unknown Nov 27 05:04:03 localhost systemd-udevd[312210]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.295 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:03.306 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '19.80.0.3/24', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=d463cc2b-c33c-4d8e-8d92-b80cf5ac63e7, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=7d36709c-2fe6-487c-9486-157a966792a0) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:03 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:03.308 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 7d36709c-2fe6-487c-9486-157a966792a0 in datapath 14034b34-9875-46a0-b548-1a34b924b9b9 bound to our chassis#033[00m Nov 27 05:04:03 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:03.314 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 8429fb01-77a6-4ffd-ac06-c71e08b7dcf0 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:04:03 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:03.314 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 14034b34-9875-46a0-b548-1a34b924b9b9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:03 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:03.315 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[cc5ce9b5-8094-4792-ac3c-c85acd302c83]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:03 localhost journal[232028]: libvirt version: 11.9.0, package: 1.el9 (builder@centos.org, 2025-11-04-09:54:50, ) Nov 27 05:04:03 localhost journal[232028]: hostname: np0005537446.localdomain Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00091|binding|INFO|Setting lport 7d36709c-2fe6-487c-9486-157a966792a0 ovn-installed in OVS Nov 27 05:04:03 localhost ovn_controller[156436]: 2025-11-27T10:04:03Z|00092|binding|INFO|Setting lport 7d36709c-2fe6-487c-9486-157a966792a0 up in Southbound Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost journal[232028]: ethtool ioctl error on tap7d36709c-2f: No such device Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.366 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.395 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e92 e92: 6 total, 6 up, 6 in Nov 27 05:04:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:03.851 2 INFO neutron.agent.securitygroups_rpc [None req-18aee0a5-9f82-4250-8016-4cb3a2040025 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Security group member updated ['6d11bb1f-93da-41d7-8007-c66992d17e28']#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.894 284030 DEBUG nova.compute.manager [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.895 284030 DEBUG oslo_concurrency.lockutils [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.895 284030 DEBUG oslo_concurrency.lockutils [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.896 284030 DEBUG oslo_concurrency.lockutils [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.896 284030 DEBUG nova.compute.manager [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] No waiting events found dispatching network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:03 localhost nova_compute[284026]: 2025-11-27 10:04:03.897 284030 WARNING nova.compute.manager [req-e0ddb1d0-0356-40d8-844c-4cfea97aa3bf req-776c52af-0675-4ff4-bd70-72561e0e1eb9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received unexpected event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for instance with vm_state active and task_state None.#033[00m Nov 27 05:04:04 localhost podman[312281]: Nov 27 05:04:04 localhost podman[312281]: 2025-11-27 10:04:04.28293967 +0000 UTC m=+0.143930579 container create 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:04 localhost podman[312281]: 2025-11-27 10:04:04.205489142 +0000 UTC m=+0.066480121 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:04:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:04:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:04:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:04:04 localhost systemd[1]: Stopping User Manager for UID 42436... Nov 27 05:04:04 localhost systemd[311769]: Activating special unit Exit the Session... Nov 27 05:04:04 localhost systemd[311769]: Stopped target Main User Target. Nov 27 05:04:04 localhost systemd[311769]: Stopped target Basic System. Nov 27 05:04:04 localhost systemd[311769]: Stopped target Paths. Nov 27 05:04:04 localhost systemd[311769]: Stopped target Sockets. Nov 27 05:04:04 localhost systemd[311769]: Stopped target Timers. Nov 27 05:04:04 localhost systemd[311769]: Stopped Mark boot as successful after the user session has run 2 minutes. Nov 27 05:04:04 localhost systemd[311769]: Stopped Daily Cleanup of User's Temporary Directories. Nov 27 05:04:04 localhost systemd[311769]: Closed D-Bus User Message Bus Socket. Nov 27 05:04:04 localhost systemd[311769]: Stopped Create User's Volatile Files and Directories. Nov 27 05:04:04 localhost systemd[311769]: Removed slice User Application Slice. Nov 27 05:04:04 localhost systemd[311769]: Reached target Shutdown. Nov 27 05:04:04 localhost systemd[311769]: Finished Exit the Session. Nov 27 05:04:04 localhost systemd[311769]: Reached target Exit the Session. Nov 27 05:04:04 localhost systemd[1]: Started libpod-conmon-40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0.scope. Nov 27 05:04:04 localhost systemd[1]: Started libcrun container. Nov 27 05:04:04 localhost systemd[1]: user@42436.service: Deactivated successfully. Nov 27 05:04:04 localhost systemd[1]: Stopped User Manager for UID 42436. Nov 27 05:04:04 localhost systemd[1]: Stopping User Runtime Directory /run/user/42436... Nov 27 05:04:04 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f4836a40e4a7d6eec8090c62b45f3fa53f77dab15c629bec5b2838dad003dfdb/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:04:04 localhost systemd[1]: run-user-42436.mount: Deactivated successfully. Nov 27 05:04:04 localhost podman[312281]: 2025-11-27 10:04:04.37668392 +0000 UTC m=+0.237674839 container init 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:04 localhost systemd[1]: user-runtime-dir@42436.service: Deactivated successfully. Nov 27 05:04:04 localhost systemd[1]: Stopped User Runtime Directory /run/user/42436. Nov 27 05:04:04 localhost systemd[1]: Removed slice User Slice of UID 42436. Nov 27 05:04:04 localhost dnsmasq[312330]: started, version 2.85 cachesize 150 Nov 27 05:04:04 localhost dnsmasq[312330]: DNS service limited to local subnets Nov 27 05:04:04 localhost podman[312281]: 2025-11-27 10:04:04.394311967 +0000 UTC m=+0.255302856 container start 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:04:04 localhost dnsmasq[312330]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:04:04 localhost dnsmasq[312330]: warning: no upstream servers configured Nov 27 05:04:04 localhost dnsmasq-dhcp[312330]: DHCP, static leases only on 19.80.0.0, lease time 1d Nov 27 05:04:04 localhost dnsmasq[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/addn_hosts - 0 addresses Nov 27 05:04:04 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/host Nov 27 05:04:04 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/opts Nov 27 05:04:04 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:04.447 265123 INFO neutron.agent.dhcp.agent [None req-813e6097-e6c8-4310-8e3a-9cc8e5a0f6c0 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:04:03Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=2161daa9-fc37-4b10-97d3-5c8dd774b7d3, ip_allocation=immediate, mac_address=fa:16:3e:69:7c:e1, name=tempest-subport-463929020, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:03:55Z, description=, dns_domain=, id=14034b34-9875-46a0-b548-1a34b924b9b9, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-subport_net-1455796487, port_security_enabled=True, project_id=a6d84e90a5fa4f74afa4103832d90ac2, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=9679, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=616, status=ACTIVE, subnets=['4d0f11d8-6923-47c1-80eb-0c3c3a33a901'], tags=[], tenant_id=a6d84e90a5fa4f74afa4103832d90ac2, updated_at=2025-11-27T10:03:58Z, vlan_transparent=None, network_id=14034b34-9875-46a0-b548-1a34b924b9b9, port_security_enabled=True, project_id=a6d84e90a5fa4f74afa4103832d90ac2, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['6d11bb1f-93da-41d7-8007-c66992d17e28'], standard_attr_id=647, status=DOWN, tags=[], tenant_id=a6d84e90a5fa4f74afa4103832d90ac2, updated_at=2025-11-27T10:04:03Z on network 14034b34-9875-46a0-b548-1a34b924b9b9#033[00m Nov 27 05:04:04 localhost podman[312297]: 2025-11-27 10:04:04.486728481 +0000 UTC m=+0.153593612 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.expose-services=, container_name=openstack_network_exporter, io.openshift.tags=minimal rhel9, distribution-scope=public, managed_by=edpm_ansible, release=1755695350, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, config_id=edpm, vcs-type=git, version=9.6) Nov 27 05:04:04 localhost podman[312295]: 2025-11-27 10:04:04.460182231 +0000 UTC m=+0.125949083 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:04:04 localhost podman[312296]: 2025-11-27 10:04:04.530942579 +0000 UTC m=+0.199804524 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:04:04 localhost podman[312297]: 2025-11-27 10:04:04.547088335 +0000 UTC m=+0.213953526 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter, name=ubi9-minimal, config_id=edpm, io.openshift.tags=minimal rhel9, architecture=x86_64, build-date=2025-08-20T13:12:41, vendor=Red Hat, Inc., vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, io.openshift.expose-services=, maintainer=Red Hat, Inc., io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, distribution-scope=public, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, com.redhat.component=ubi9-minimal-container) Nov 27 05:04:04 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:04:04 localhost podman[312296]: 2025-11-27 10:04:04.568932427 +0000 UTC m=+0.237794382 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:04:04 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:04:04 localhost podman[312295]: 2025-11-27 10:04:04.602333352 +0000 UTC m=+0.268100244 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125) Nov 27 05:04:04 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:04.605 265123 INFO neutron.agent.dhcp.agent [None req-cfbf98a7-ebad-4952-9820-8c69b253087a - - - - - -] DHCP configuration for ports {'0b3b33c8-3d8e-4978-ae70-eedba54823eb'} is completed#033[00m Nov 27 05:04:04 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:04:04 localhost dnsmasq[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/addn_hosts - 1 addresses Nov 27 05:04:04 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/host Nov 27 05:04:04 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/opts Nov 27 05:04:04 localhost podman[312380]: 2025-11-27 10:04:04.756820696 +0000 UTC m=+0.065496304 container kill 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:04:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v137: 177 pgs: 177 active+clean; 304 MiB data, 1008 MiB used, 41 GiB / 42 GiB avail; 547 KiB/s rd, 2.7 MiB/s wr, 106 op/s Nov 27 05:04:05 localhost systemd[1]: tmp-crun.o19o4j.mount: Deactivated successfully. Nov 27 05:04:05 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:05.381 265123 INFO neutron.agent.dhcp.agent [None req-dec3e8af-34b3-40d6-befc-26b7549c77e2 - - - - - -] DHCP configuration for ports {'2161daa9-fc37-4b10-97d3-5c8dd774b7d3'} is completed#033[00m Nov 27 05:04:05 localhost nova_compute[284026]: 2025-11-27 10:04:05.476 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:05 localhost ovn_controller[156436]: 2025-11-27T10:04:05Z|00093|binding|INFO|Releasing lport cdeccc19-a6f6-4ac9-b28c-bcd5b506ddbe from this chassis (sb_readonly=0) Nov 27 05:04:05 localhost ovn_controller[156436]: 2025-11-27T10:04:05Z|00094|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:05 localhost ovn_controller[156436]: 2025-11-27T10:04:05Z|00095|binding|INFO|Releasing lport b5e8fcfd-584f-4244-9345-c50b7c0b100b from this chassis (sb_readonly=0) Nov 27 05:04:05 localhost podman[312418]: 2025-11-27 10:04:05.621233211 +0000 UTC m=+0.064214510 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:04:05 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:04:05 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:05 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:05 localhost nova_compute[284026]: 2025-11-27 10:04:05.650 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e93 e93: 6 total, 6 up, 6 in Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.294 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.295 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.312 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Starting instance... _do_build_and_run_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:2402#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.402 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.403 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.410 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Require both a host and instance NUMA topology to fit instance on host. numa_fit_instance_to_host /usr/lib/python3.9/site-packages/nova/virt/hardware.py:2368#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.411 284030 INFO nova.compute.claims [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Claim successful on node np0005537446.localdomain#033[00m Nov 27 05:04:06 localhost nova_compute[284026]: 2025-11-27 10:04:06.574 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e94 e94: 6 total, 6 up, 6 in Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.098 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.524s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v140: 177 pgs: 177 active+clean; 383 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 7.8 MiB/s rd, 7.8 MiB/s wr, 138 op/s Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.109 284030 DEBUG nova.compute.provider_tree [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.126 284030 DEBUG nova.scheduler.client.report [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.152 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.749s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.154 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Start building networks asynchronously for instance. _build_resources /usr/lib/python3.9/site-packages/nova/compute/manager.py:2799#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.231 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Allocating IP information in the background. _allocate_network_async /usr/lib/python3.9/site-packages/nova/compute/manager.py:1952#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.232 284030 DEBUG nova.network.neutron [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] allocate_for_instance() allocate_for_instance /usr/lib/python3.9/site-packages/nova/network/neutron.py:1156#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.254 284030 INFO nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Ignoring supplied device name: /dev/vda. Libvirt can't honour user-supplied dev names#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.274 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Start building block device mappings for instance. _build_resources /usr/lib/python3.9/site-packages/nova/compute/manager.py:2834#033[00m Nov 27 05:04:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e94 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.406 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Start spawning the instance on the hypervisor. _build_and_run_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:2608#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.408 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Creating instance directory _create_image /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4723#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.409 284030 INFO nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Creating image(s)#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.452 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.494 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.534 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.539 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "dbc4c9cab63f62d6f95108aba598490d75e1fc03" by "nova.virt.libvirt.imagebackend.Image.cache..fetch_func_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.541 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "dbc4c9cab63f62d6f95108aba598490d75e1fc03" acquired by "nova.virt.libvirt.imagebackend.Image.cache..fetch_func_sync" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.547 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.575 284030 DEBUG nova.virt.libvirt.imagebackend [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Image locations are: [{'url': 'rbd://e83f3b0c-4090-52df-95d4-ad9be8516692/images/a85008d0-4cdf-44ee-935d-c55441a5aa59/snap', 'metadata': {'store': 'default_backend'}}, {'url': 'rbd://e83f3b0c-4090-52df-95d4-ad9be8516692/images/a85008d0-4cdf-44ee-935d-c55441a5aa59/snap', 'metadata': {}}] clone /usr/lib/python3.9/site-packages/nova/virt/libvirt/imagebackend.py:1085#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.721 284030 WARNING oslo_policy.policy [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] JSON formatted policy_file support is deprecated since Victoria release. You need to use YAML format which will be default in future. You can use ``oslopolicy-convert-json-to-yaml`` tool to convert existing JSON-formatted policy file to YAML-formatted in backward compatible way: https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html.#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.722 284030 WARNING oslo_policy.policy [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] JSON formatted policy_file support is deprecated since Victoria release. You need to use YAML format which will be default in future. You can use ``oslopolicy-convert-json-to-yaml`` tool to convert existing JSON-formatted policy file to YAML-formatted in backward compatible way: https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html.#033[00m Nov 27 05:04:07 localhost nova_compute[284026]: 2025-11-27 10:04:07.727 284030 DEBUG nova.policy [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'edb03acaf09647f6b6b5465b16878a77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} authorize /usr/lib/python3.9/site-packages/nova/policy.py:203#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.427 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): /usr/bin/python3 -m oslo_concurrency.prlimit --as=1073741824 --cpu=30 -- env LC_ALL=C LANG=C qemu-img info /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.part --force-share --output=json execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.512 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "/usr/bin/python3 -m oslo_concurrency.prlimit --as=1073741824 --cpu=30 -- env LC_ALL=C LANG=C qemu-img info /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.part --force-share --output=json" returned: 0 in 0.085s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.513 284030 DEBUG nova.virt.images [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] a85008d0-4cdf-44ee-935d-c55441a5aa59 was qcow2, converting to raw fetch_to_raw /usr/lib/python3.9/site-packages/nova/virt/images.py:242#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.515 284030 DEBUG nova.privsep.utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Path '/var/lib/nova/instances' supports direct I/O supports_direct_io /usr/lib/python3.9/site-packages/nova/privsep/utils.py:63#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.516 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): qemu-img convert -t none -O raw -f qcow2 /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.part /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.converted execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.580 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.580 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.581 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.581 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.581 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.582 284030 INFO nova.compute.manager [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Terminating instance#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.584 284030 DEBUG nova.compute.manager [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Start destroying the instance on the hypervisor. _shutdown_instance /usr/lib/python3.9/site-packages/nova/compute/manager.py:3120#033[00m Nov 27 05:04:08 localhost kernel: device tap30bcf056-09 left promiscuous mode Nov 27 05:04:08 localhost NetworkManager[5971]: [1764237848.7031] device (tap30bcf056-09): state change: disconnected -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.711 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00096|binding|INFO|Releasing lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 from this chassis (sb_readonly=0) Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00097|binding|INFO|Setting lport 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 down in Southbound Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00098|binding|INFO|Releasing lport 86dea736-4b08-40ac-86cd-1432d75dbf91 from this chassis (sb_readonly=0) Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00099|binding|INFO|Setting lport 86dea736-4b08-40ac-86cd-1432d75dbf91 down in Southbound Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00100|binding|INFO|Removing iface tap30bcf056-09 ovn-installed in OVS Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.717 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.721 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:ed:d8:4f 19.80.0.59'], port_security=['fa:16:3e:ed:d8:4f 19.80.0.59'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=['30bcf056-09f4-4ccc-94c7-ca6d81a754e9'], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-subport-447612827', 'neutron:cidrs': '19.80.0.59/24', 'neutron:device_id': '', 'neutron:device_owner': 'trunk:subport', 'neutron:mtu': '', 'neutron:network_name': 'neutron-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-subport-447612827', 'neutron:project_id': '0780cc2b55f74fbd8c97a1f5f39cd8cc', 'neutron:revision_number': '5', 'neutron:security_group_ids': 'd5a4dbad-2a9b-45ce-81ca-0430479e3c01', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[42], additional_encap=[], encap=[], mirror_rules=[], datapath=174951dd-7028-48c6-9b6e-de8f9b584ff5, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=86dea736-4b08-40ac-86cd-1432d75dbf91) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00101|binding|INFO|Releasing lport cdeccc19-a6f6-4ac9-b28c-bcd5b506ddbe from this chassis (sb_readonly=0) Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00102|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:08 localhost ovn_controller[156436]: 2025-11-27T10:04:08Z|00103|binding|INFO|Releasing lport b5e8fcfd-584f-4244-9345-c50b7c0b100b from this chassis (sb_readonly=0) Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.723 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:83:7f:f2 10.100.0.7'], port_security=['fa:16:3e:83:7f:f2 10.100.0.7'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-parent-1295767349', 'neutron:cidrs': '10.100.0.7/28', 'neutron:device_id': 'ea7c65ed-7232-403d-84e6-2ead60aade7c', 'neutron:device_owner': 'compute:nova', 'neutron:mtu': '', 'neutron:network_name': 'neutron-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-parent-1295767349', 'neutron:project_id': '0780cc2b55f74fbd8c97a1f5f39cd8cc', 'neutron:revision_number': '11', 'neutron:security_group_ids': 'd5a4dbad-2a9b-45ce-81ca-0430479e3c01', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=1cc89ba8-681c-4c44-a43f-cbe7c9b9d281, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=30bcf056-09f4-4ccc-94c7-ca6d81a754e9) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.724 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 86dea736-4b08-40ac-86cd-1432d75dbf91 in datapath 89c93072-e57c-4d59-b8c2-3bb1a82cf80c unbound from our chassis#033[00m Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.727 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 89c93072-e57c-4d59-b8c2-3bb1a82cf80c, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.731 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[f6556302-c628-40fe-bc87-b260040698ec]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:08.731 162092 INFO neutron.agent.ovn.metadata.agent [-] Cleaning up ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c namespace which is not needed anymore#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.748 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "qemu-img convert -t none -O raw -f qcow2 /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.part /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.converted" returned: 0 in 0.232s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:08 localhost systemd[1]: machine-qemu\x2d3\x2dinstance\x2d00000006.scope: Deactivated successfully. Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.752 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): /usr/bin/python3 -m oslo_concurrency.prlimit --as=1073741824 --cpu=30 -- env LC_ALL=C LANG=C qemu-img info /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.converted --force-share --output=json execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:08 localhost systemd[1]: machine-qemu\x2d3\x2dinstance\x2d00000006.scope: Consumed 1.371s CPU time. Nov 27 05:04:08 localhost systemd-machined[86026]: Machine qemu-3-instance-00000006 terminated. Nov 27 05:04:08 localhost podman[242678]: time="2025-11-27T10:04:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.784 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e95 e95: 6 total, 6 up, 6 in Nov 27 05:04:08 localhost podman[242678]: @ - - [27/Nov/2025:10:04:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 163532 "" "Go-http-client/1.1" Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.831 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "/usr/bin/python3 -m oslo_concurrency.prlimit --as=1073741824 --cpu=30 -- env LC_ALL=C LANG=C qemu-img info /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03.converted --force-share --output=json" returned: 0 in 0.079s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.832 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "dbc4c9cab63f62d6f95108aba598490d75e1fc03" "released" by "nova.virt.libvirt.imagebackend.Image.cache..fetch_func_sync" :: held 1.291s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:08 localhost podman[242678]: @ - - [27/Nov/2025:10:04:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 21642 "" "Go-http-client/1.1" Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.862 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.866 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): rbd import --pool vms /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk --image-format=2 --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.885 284030 INFO nova.virt.libvirt.driver [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Instance destroyed successfully.#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.886 284030 DEBUG nova.objects.instance [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lazy-loading 'resources' on Instance uuid ea7c65ed-7232-403d-84e6-2ead60aade7c obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.901 284030 DEBUG nova.virt.libvirt.vif [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=True,config_drive='True',created_at=2025-11-27T10:03:11Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description=None,display_name='tempest-LiveAutoBlockMigrationV225Test-server-1085268656',ec2_ids=,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537446.localdomain',hostname='tempest-liveautoblockmigrationv225test-server-1085268656',id=6,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T10:03:37Z,launched_on='np0005537444.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=,power_state=1,progress=0,project_id='0780cc2b55f74fbd8c97a1f5f39cd8cc',ramdisk_id='',reservation_id='r-f7ny7m85',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',clean_attempts='1',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='q35',image_hw_pointer_model='usbtablet',image_hw_rng_model='virtio',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',owner_project_name='tempest-LiveAutoBlockMigrationV225Test-1706287954',owner_user_name='tempest-LiveAutoBlockMigrationV225Test-1706287954-project-member'},tags=,task_state='deleting',terminated_at=None,trusted_certs=,updated_at=2025-11-27T10:03:58Z,user_data=None,user_id='730775ba58804b5b8977106a8a8aa97c',uuid=ea7c65ed-7232-403d-84e6-2ead60aade7c,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} unplug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:828#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.901 284030 DEBUG nova.network.os_vif_util [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Converting VIF {"id": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "address": "fa:16:3e:83:7f:f2", "network": {"id": "f966841f-0fe6-48fb-bc34-0c4a63a8d9fc", "bridge": "br-int", "label": "tempest-LiveAutoBlockMigrationV225Test-775107190-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.2"}}], "meta": {"injected": false, "tenant_id": "0780cc2b55f74fbd8c97a1f5f39cd8cc", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap30bcf056-09", "ovs_interfaceid": "30bcf056-09f4-4ccc-94c7-ca6d81a754e9", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.901 284030 DEBUG nova.network.os_vif_util [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.902 284030 DEBUG os_vif [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Unplugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09') unplug /usr/lib/python3.9/site-packages/os_vif/__init__.py:109#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.904 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.904 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap30bcf056-09, bridge=br-int, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.906 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.907 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.910 284030 INFO os_vif [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Successfully unplugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:83:7f:f2,bridge_name='br-int',has_traffic_filtering=True,id=30bcf056-09f4-4ccc-94c7-ca6d81a754e9,network=Network(f966841f-0fe6-48fb-bc34-0c4a63a8d9fc),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap30bcf056-09')#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.993 284030 DEBUG nova.compute.manager [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-unplugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.994 284030 DEBUG oslo_concurrency.lockutils [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.994 284030 DEBUG oslo_concurrency.lockutils [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.995 284030 DEBUG oslo_concurrency.lockutils [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.995 284030 DEBUG nova.compute.manager [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] No waiting events found dispatching network-vif-unplugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:08 localhost nova_compute[284026]: 2025-11-27 10:04:08.995 284030 DEBUG nova.compute.manager [req-5629af5f-4a51-44b3-82e0-3c322724ba3c req-adba83f4-2db4-4978-803a-2ecad7225cf9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-unplugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for instance with task_state deleting. _process_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:10826#033[00m Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [NOTICE] (311990) : haproxy version is 2.8.14-c23fe91 Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [NOTICE] (311990) : path to executable is /usr/sbin/haproxy Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [WARNING] (311990) : Exiting Master process... Nov 27 05:04:09 localhost systemd[1]: tmp-crun.9hpEcM.mount: Deactivated successfully. Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [WARNING] (311990) : Exiting Master process... Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [ALERT] (311990) : Current worker (312000) exited with code 143 (Terminated) Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c[311986]: [WARNING] (311990) : All workers exited. Exiting... (0) Nov 27 05:04:09 localhost systemd[1]: libpod-a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5.scope: Deactivated successfully. Nov 27 05:04:09 localhost podman[312582]: 2025-11-27 10:04:09.02094586 +0000 UTC m=+0.095498068 container died a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:09 localhost podman[312582]: 2025-11-27 10:04:09.082010004 +0000 UTC m=+0.156562192 container cleanup a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:04:09 localhost podman[312628]: 2025-11-27 10:04:09.093094264 +0000 UTC m=+0.065417862 container cleanup a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:04:09 localhost systemd[1]: libpod-conmon-a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5.scope: Deactivated successfully. Nov 27 05:04:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v142: 177 pgs: 177 active+clean; 383 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 8.7 MiB/s rd, 8.6 MiB/s wr, 153 op/s Nov 27 05:04:09 localhost podman[312647]: 2025-11-27 10:04:09.173343688 +0000 UTC m=+0.076358249 container remove a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.178 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[149a7cbe-4e52-402d-893a-a68b6c3558d0]: (4, ('Thu Nov 27 10:04:08 AM UTC 2025 Stopping container neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c (a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5)\na0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5\nThu Nov 27 10:04:09 AM UTC 2025 Deleting container neutron-haproxy-ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c (a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5)\na0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.180 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b2c4d806-9ef5-46c5-851e-7d50db8b4bcd]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.182 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap89c93072-e0, bridge=None, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.184 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:09 localhost kernel: device tap89c93072-e0 left promiscuous mode Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.192 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.201 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4597d3c4-119c-4c95-86a4-a9a863db6c9f]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.221 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a370d54d-a67a-4d97-8099-c3b14f3ac9ec]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.222 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[57e49bbd-a177-4b23-be5c-a72f019f998a]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.235 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[66a24396-cfde-4dad-8291-4718a45a32b8]: (4, [{'family': 0, '__align': (), 'ifi_type': 772, 'index': 1, 'flags': 65609, 'change': 0, 'attrs': [['IFLA_IFNAME', 'lo'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UNKNOWN'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 65536], ['IFLA_MIN_MTU', 0], ['IFLA_MAX_MTU', 0], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 1], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 1], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 0], ['IFLA_CARRIER_UP_COUNT', 0], ['IFLA_CARRIER_DOWN_COUNT', 0], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', '00:00:00:00:00:00'], ['IFLA_BROADCAST', '00:00:00:00:00:00'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 1, 'nopolicy': 1, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206523, 'reachable_time': 27206, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 65536, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 4294967295, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 4294967295, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 0, 'inoctets': 0, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 0, 'outoctets': 0, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 0, 'outmcastpkts': 0, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 0, 'outmcastoctets': 0, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 0, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 0, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1356, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 312661, 'error': None, 'target': 'ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.237 162227 DEBUG neutron.privileged.agent.linux.ip_lib [-] Namespace ovnmeta-89c93072-e57c-4d59-b8c2-3bb1a82cf80c deleted. remove_netns /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:607#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.237 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[dd46a7e2-aa19-49fd-8571-d8e17e86e252]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.238 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 30bcf056-09f4-4ccc-94c7-ca6d81a754e9 in datapath f966841f-0fe6-48fb-bc34-0c4a63a8d9fc unbound from our chassis#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.241 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.242 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4e41d382-c7f6-4d01-b9f5-78350e75e109]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.243 162092 INFO neutron.agent.ovn.metadata.agent [-] Cleaning up ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc namespace which is not needed anymore#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.278 284030 DEBUG nova.network.neutron [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Successfully updated port: 20252d49-04fc-4922-8fc1-99b7e3b77a67 _update_port /usr/lib/python3.9/site-packages/nova/network/neutron.py:586#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.307 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.307 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquired lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.308 284030 DEBUG nova.network.neutron [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Building network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2010#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.384 284030 DEBUG nova.network.neutron [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Instance cache missing network info. _get_preexisting_port_ids /usr/lib/python3.9/site-packages/nova/network/neutron.py:3323#033[00m Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [NOTICE] (312132) : haproxy version is 2.8.14-c23fe91 Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [NOTICE] (312132) : path to executable is /usr/sbin/haproxy Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [WARNING] (312132) : Exiting Master process... Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [WARNING] (312132) : Exiting Master process... Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [ALERT] (312132) : Current worker (312134) exited with code 143 (Terminated) Nov 27 05:04:09 localhost neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc[312120]: [WARNING] (312132) : All workers exited. Exiting... (0) Nov 27 05:04:09 localhost systemd[1]: libpod-3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed.scope: Deactivated successfully. Nov 27 05:04:09 localhost podman[312679]: 2025-11-27 10:04:09.422497037 +0000 UTC m=+0.077118710 container died 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.431 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "rbd import --pool vms /var/lib/nova/instances/_base/dbc4c9cab63f62d6f95108aba598490d75e1fc03 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk --image-format=2 --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.564s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:09 localhost podman[312679]: 2025-11-27 10:04:09.454884934 +0000 UTC m=+0.109506597 container cleanup 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:04:09 localhost podman[312693]: 2025-11-27 10:04:09.503160743 +0000 UTC m=+0.070822781 container cleanup 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:09 localhost systemd[1]: libpod-conmon-3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed.scope: Deactivated successfully. Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.539 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] resizing rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk to 1073741824 resize /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:288#033[00m Nov 27 05:04:09 localhost podman[312723]: 2025-11-27 10:04:09.551396509 +0000 UTC m=+0.073270936 container remove 3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.556 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[07903dff-f215-4029-8111-41c837317e3e]: (4, ('Thu Nov 27 10:04:09 AM UTC 2025 Stopping container neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc (3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed)\n3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed\nThu Nov 27 10:04:09 AM UTC 2025 Deleting container neutron-haproxy-ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc (3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed)\n3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.558 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[9e5deaee-3e83-49fa-b3a8-594c4963aac8]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.559 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tapf966841f-00, bridge=None, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:09 localhost kernel: device tapf966841f-00 left promiscuous mode Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.569 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ef7853b7-1d72-4c72-bc0f-1c3eb8ff9abf]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.587 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[670a85ea-4f0f-4c9d-b7f4-9dd57ea65686]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.588 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d7b9f56b-fef0-4cec-98a4-9a586a4137db]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.602 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.604 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ac318107-7b93-4906-8b95-c662e7fac2f5]: (4, [{'family': 0, '__align': (), 'ifi_type': 772, 'index': 1, 'flags': 65609, 'change': 0, 'attrs': [['IFLA_IFNAME', 'lo'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UNKNOWN'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 65536], ['IFLA_MIN_MTU', 0], ['IFLA_MAX_MTU', 0], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 1], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 1], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 0], ['IFLA_CARRIER_UP_COUNT', 0], ['IFLA_CARRIER_DOWN_COUNT', 0], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', '00:00:00:00:00:00'], ['IFLA_BROADCAST', '00:00:00:00:00:00'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 1, 'nopolicy': 1, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1206615, 'reachable_time': 40222, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 65536, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 4294967295, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 4294967295, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 0, 'inoctets': 0, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 0, 'outoctets': 0, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 0, 'outmcastpkts': 0, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 0, 'outmcastoctets': 0, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 0, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 0, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1356, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 312778, 'error': None, 'target': 'ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.606 162227 DEBUG neutron.privileged.agent.linux.ip_lib [-] Namespace ovnmeta-f966841f-0fe6-48fb-bc34-0c4a63a8d9fc deleted. remove_netns /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:607#033[00m Nov 27 05:04:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:09.606 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[0ff60b16-4486-4e1c-b2f5-78f66f27dbca]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.729 284030 DEBUG nova.objects.instance [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lazy-loading 'migration_context' on Instance uuid 765bb3f2-7b81-4d43-8eea-4f8b0e498677 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.754 284030 INFO nova.virt.libvirt.driver [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Deleting instance files /var/lib/nova/instances/ea7c65ed-7232-403d-84e6-2ead60aade7c_del#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.755 284030 INFO nova.virt.libvirt.driver [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Deletion of /var/lib/nova/instances/ea7c65ed-7232-403d-84e6-2ead60aade7c_del complete#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.759 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Created local disks _create_image /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4857#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.760 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Ensure instance console log exists: /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/console.log _ensure_console_log_for_instance /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4609#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.760 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "vgpu_resources" by "nova.virt.libvirt.driver.LibvirtDriver._allocate_mdevs" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.760 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "vgpu_resources" acquired by "nova.virt.libvirt.driver.LibvirtDriver._allocate_mdevs" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.761 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "vgpu_resources" "released" by "nova.virt.libvirt.driver.LibvirtDriver._allocate_mdevs" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e96 e96: 6 total, 6 up, 6 in Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.829 284030 DEBUG nova.network.neutron [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updating instance_info_cache with network_info: [{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.859 284030 INFO nova.compute.manager [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Took 1.27 seconds to destroy the instance on the hypervisor.#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.860 284030 DEBUG oslo.service.loopingcall [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. func /usr/lib/python3.9/site-packages/oslo_service/loopingcall.py:435#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.861 284030 DEBUG nova.compute.manager [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Deallocating network for instance _deallocate_network /usr/lib/python3.9/site-packages/nova/compute/manager.py:2259#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.861 284030 DEBUG nova.network.neutron [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] deallocate_for_instance() deallocate_for_instance /usr/lib/python3.9/site-packages/nova/network/neutron.py:1803#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.870 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Releasing lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.871 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Instance network_info: |[{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}]| _allocate_network_async /usr/lib/python3.9/site-packages/nova/compute/manager.py:1967#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.875 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Start _get_guest_xml network_info=[{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] disk_info={'disk_bus': 'virtio', 'cdrom_bus': 'sata', 'mapping': {'root': {'bus': 'virtio', 'dev': 'vda', 'type': 'disk', 'boot_index': '1'}, 'disk': {'bus': 'virtio', 'dev': 'vda', 'type': 'disk', 'boot_index': '1'}, 'disk.config': {'bus': 'sata', 'dev': 'sda', 'type': 'cdrom'}}} image_meta=ImageMeta(checksum='c8fc807773e5354afe61636071771906',container_format='bare',created_at=2025-11-27T10:01:18Z,direct_url=,disk_format='qcow2',id=a85008d0-4cdf-44ee-935d-c55441a5aa59,min_disk=0,min_ram=0,name='cirros-0.6.2-x86_64-disk.img',owner='02a308de13ca4518beaddd402ba93936',properties=ImageMetaProps,protected=,size=21430272,status='active',tags=,updated_at=2025-11-27T10:01:21Z,virtual_size=,visibility=) rescue=None block_device_info={'root_device_name': '/dev/vda', 'image': [{'boot_index': 0, 'device_name': '/dev/vda', 'encrypted': False, 'guest_format': None, 'size': 0, 'encryption_format': None, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': 'virtio', 'encryption_secret_uuid': None, 'image_id': 'a85008d0-4cdf-44ee-935d-c55441a5aa59'}], 'ephemerals': [], 'block_device_mapping': [], 'swap': None} _get_guest_xml /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:7549#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.881 284030 WARNING nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.885 284030 DEBUG nova.virt.libvirt.host [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Searching host: 'np0005537446.localdomain' for CPU controller through CGroups V1... _has_cgroupsv1_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1653#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.886 284030 DEBUG nova.virt.libvirt.host [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CPU controller missing on host. _has_cgroupsv1_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1663#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.888 284030 DEBUG nova.virt.libvirt.host [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Searching host: 'np0005537446.localdomain' for CPU controller through CGroups V2... _has_cgroupsv2_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1672#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.888 284030 DEBUG nova.virt.libvirt.host [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CPU controller found on host. _has_cgroupsv2_cpu_controller /usr/lib/python3.9/site-packages/nova/virt/libvirt/host.py:1679#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.889 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CPU mode 'host-model' models '' was chosen, with extra flags: '' _get_guest_cpu_model_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:5396#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.889 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Getting desirable topologies for flavor Flavor(created_at=2025-11-27T10:01:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='88120525-2aa5-45cc-ae1c-52b1162513c1',id=5,is_public=True,memory_mb=128,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c8fc807773e5354afe61636071771906',container_format='bare',created_at=2025-11-27T10:01:18Z,direct_url=,disk_format='qcow2',id=a85008d0-4cdf-44ee-935d-c55441a5aa59,min_disk=0,min_ram=0,name='cirros-0.6.2-x86_64-disk.img',owner='02a308de13ca4518beaddd402ba93936',properties=ImageMetaProps,protected=,size=21430272,status='active',tags=,updated_at=2025-11-27T10:01:21Z,virtual_size=,visibility=), allow threads: True _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:563#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.889 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Flavor limits 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:348#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.890 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Image limits 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:352#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.890 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Flavor pref 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:388#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.890 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Image pref 0:0:0 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:392#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.890 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 get_cpu_topology_constraints /usr/lib/python3.9/site-packages/nova/virt/hardware.py:430#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.890 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:569#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.891 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Build topologies for 1 vcpu(s) 1:1:1 _get_possible_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:471#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.891 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Got 1 possible topologies _get_possible_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:501#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.891 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:575#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.892 284030 DEBUG nova.virt.hardware [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] _get_desirable_cpu_topologies /usr/lib/python3.9/site-packages/nova/virt/hardware.py:577#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.898 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.975 284030 DEBUG nova.compute.manager [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-changed-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.975 284030 DEBUG nova.compute.manager [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Refreshing instance network info cache due to event network-changed-20252d49-04fc-4922-8fc1-99b7e3b77a67. external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11053#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.976 284030 DEBUG oslo_concurrency.lockutils [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.976 284030 DEBUG oslo_concurrency.lockutils [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquired lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:04:09 localhost nova_compute[284026]: 2025-11-27 10:04:09.977 284030 DEBUG nova.network.neutron [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Refreshing network info cache for port 20252d49-04fc-4922-8fc1-99b7e3b77a67 _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2007#033[00m Nov 27 05:04:10 localhost systemd[1]: var-lib-containers-storage-overlay-fe2810b887e4c6ff976b50fdb48de78b18e4541fb2a46d627acc143cdfdba828-merged.mount: Deactivated successfully. Nov 27 05:04:10 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-3d7939d48d1f8c3fb32599c656014e3c17db1cefab7aba6cd801d8eb3e65e8ed-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:10 localhost systemd[1]: run-netns-ovnmeta\x2df966841f\x2d0fe6\x2d48fb\x2dbc34\x2d0c4a63a8d9fc.mount: Deactivated successfully. Nov 27 05:04:10 localhost systemd[1]: var-lib-containers-storage-overlay-c888ae8a8f2d5df990e7266ba4dbb42cc2b9a0b65d0ce1dc20c878b8f1e7170b-merged.mount: Deactivated successfully. Nov 27 05:04:10 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-a0eca2561aa033169b3a1c660e66cd98924fd6f7d5c01a8c48efbba7db8498c5-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:10 localhost systemd[1]: run-netns-ovnmeta\x2d89c93072\x2de57c\x2d4d59\x2db8c2\x2d3bb1a82cf80c.mount: Deactivated successfully. Nov 27 05:04:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:04:10 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/4080495822' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.352 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.454s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.386 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk.config does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.391 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.516 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.654 284030 DEBUG nova.network.neutron [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updated VIF entry in instance network info cache for port 20252d49-04fc-4922-8fc1-99b7e3b77a67. _build_network_info_model /usr/lib/python3.9/site-packages/nova/network/neutron.py:3482#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.655 284030 DEBUG nova.network.neutron [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updating instance_info_cache with network_info: [{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.677 284030 DEBUG oslo_concurrency.lockutils [req-3056380c-3eff-4a8e-afa4-4f0db8c73989 req-53c9ad38-2ee9-42f1-be21-d693476855f4 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Releasing lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:04:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:04:10 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1734381793' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.806 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "ceph mon dump --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.416s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.808 284030 DEBUG nova.virt.libvirt.vif [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T10:04:05Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=None,disable_terminate=False,display_description='tempest-LiveMigrationTest-server-1077612390',display_name='tempest-LiveMigrationTest-server-1077612390',ec2_ids=EC2Ids,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537446.localdomain',hostname='tempest-livemigrationtest-server-1077612390',id=8,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=KeyPairList,launch_index=0,launched_at=None,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=None,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=InstancePCIRequests,power_state=0,progress=0,project_id='a6d84e90a5fa4f74afa4103832d90ac2',ramdisk_id='',reservation_id='r-mchq7ras',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_machine_type='q35',image_hw_rng_model='virtio',image_min_disk='1',image_min_ram='0',network_allocated='True',owner_project_name='tempest-LiveMigrationTest-1823599681',owner_user_name='tempest-LiveMigrationTest-1823599681-project-member'},tags=TagList,task_state='spawning',terminated_at=None,trusted_certs=None,updated_at=2025-11-27T10:04:07Z,user_data=None,user_id='edb03acaf09647f6b6b5465b16878a77',uuid=765bb3f2-7b81-4d43-8eea-4f8b0e498677,vcpu_model=VirtCPUModel,vcpus=1,vm_mode=None,vm_state='building') vif={"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} virt_type=kvm get_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:563#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.809 284030 DEBUG nova.network.os_vif_util [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Converting VIF {"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.810 284030 DEBUG nova.network.os_vif_util [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.812 284030 DEBUG nova.objects.instance [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lazy-loading 'pci_devices' on Instance uuid 765bb3f2-7b81-4d43-8eea-4f8b0e498677 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.831 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] End _get_guest_xml xml= Nov 27 05:04:10 localhost nova_compute[284026]: 765bb3f2-7b81-4d43-8eea-4f8b0e498677 Nov 27 05:04:10 localhost nova_compute[284026]: instance-00000008 Nov 27 05:04:10 localhost nova_compute[284026]: 131072 Nov 27 05:04:10 localhost nova_compute[284026]: 1 Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: tempest-LiveMigrationTest-server-1077612390 Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:09 Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: 128 Nov 27 05:04:10 localhost nova_compute[284026]: 1 Nov 27 05:04:10 localhost nova_compute[284026]: 0 Nov 27 05:04:10 localhost nova_compute[284026]: 0 Nov 27 05:04:10 localhost nova_compute[284026]: 1 Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: tempest-LiveMigrationTest-1823599681-project-member Nov 27 05:04:10 localhost nova_compute[284026]: tempest-LiveMigrationTest-1823599681 Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: RDO Nov 27 05:04:10 localhost nova_compute[284026]: OpenStack Compute Nov 27 05:04:10 localhost nova_compute[284026]: 27.5.2-0.20250829104910.6f8decf.el9 Nov 27 05:04:10 localhost nova_compute[284026]: 765bb3f2-7b81-4d43-8eea-4f8b0e498677 Nov 27 05:04:10 localhost nova_compute[284026]: 765bb3f2-7b81-4d43-8eea-4f8b0e498677 Nov 27 05:04:10 localhost nova_compute[284026]: Virtual Machine Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: hvm Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: /dev/urandom Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: Nov 27 05:04:10 localhost nova_compute[284026]: _get_guest_xml /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:7555#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.832 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Preparing to wait for external event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 prepare_for_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:283#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.833 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.prepare_for_instance_event.._create_or_get_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.833 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.prepare_for_instance_event.._create_or_get_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.833 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.prepare_for_instance_event.._create_or_get_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.834 284030 DEBUG nova.virt.libvirt.vif [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='',created_at=2025-11-27T10:04:05Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=None,disable_terminate=False,display_description='tempest-LiveMigrationTest-server-1077612390',display_name='tempest-LiveMigrationTest-server-1077612390',ec2_ids=EC2Ids,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537446.localdomain',hostname='tempest-livemigrationtest-server-1077612390',id=8,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=KeyPairList,launch_index=0,launched_at=None,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=None,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=PciDeviceList,pci_requests=InstancePCIRequests,power_state=0,progress=0,project_id='a6d84e90a5fa4f74afa4103832d90ac2',ramdisk_id='',reservation_id='r-mchq7ras',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_machine_type='q35',image_hw_rng_model='virtio',image_min_disk='1',image_min_ram='0',network_allocated='True',owner_project_name='tempest-LiveMigrationTest-1823599681',owner_user_name='tempest-LiveMigrationTest-1823599681-project-member'},tags=TagList,task_state='spawning',terminated_at=None,trusted_certs=None,updated_at=2025-11-27T10:04:07Z,user_data=None,user_id='edb03acaf09647f6b6b5465b16878a77',uuid=765bb3f2-7b81-4d43-8eea-4f8b0e498677,vcpu_model=VirtCPUModel,vcpus=1,vm_mode=None,vm_state='building') vif={"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} plug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:710#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.835 284030 DEBUG nova.network.os_vif_util [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Converting VIF {"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.835 284030 DEBUG nova.network.os_vif_util [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.836 284030 DEBUG os_vif [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Plugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') plug /usr/lib/python3.9/site-packages/os_vif/__init__.py:76#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.837 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.837 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddBridgeCommand(_result=None, name=br-int, may_exist=True, datapath_type=system) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.838 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.841 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.841 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap20252d49-04, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.842 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Interface, record=tap20252d49-04, col_values=(('external_ids', {'iface-id': '20252d49-04fc-4922-8fc1-99b7e3b77a67', 'iface-status': 'active', 'attached-mac': 'fa:16:3e:75:1f:c2', 'vm-uuid': '765bb3f2-7b81-4d43-8eea-4f8b0e498677'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.844 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.846 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.848 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.849 284030 INFO os_vif [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Successfully plugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04')#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.929 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] No BDM found with device name vda, not building metadata. _build_disk_metadata /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:12116#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.929 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] No BDM found with device name sda, not building metadata. _build_disk_metadata /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:12116#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.930 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] No VIF found with MAC fa:16:3e:75:1f:c2, not building metadata _build_interface_metadata /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:12092#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.932 284030 INFO nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Using config drive#033[00m Nov 27 05:04:10 localhost nova_compute[284026]: 2025-11-27 10:04:10.971 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk.config does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.075 284030 DEBUG nova.compute.manager [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.076 284030 DEBUG oslo_concurrency.lockutils [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.076 284030 DEBUG oslo_concurrency.lockutils [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.076 284030 DEBUG oslo_concurrency.lockutils [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.077 284030 DEBUG nova.compute.manager [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] No waiting events found dispatching network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.077 284030 WARNING nova.compute.manager [req-d13caf8f-93be-4801-9604-a51d5a74b4e1 req-9167b43a-ce05-4c61-9c5b-3ef2354a09c6 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Received unexpected event network-vif-plugged-30bcf056-09f4-4ccc-94c7-ca6d81a754e9 for instance with vm_state active and task_state deleting.#033[00m Nov 27 05:04:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v144: 177 pgs: 177 active+clean; 271 MiB data, 970 MiB used, 41 GiB / 42 GiB avail; 4.0 MiB/s rd, 4.0 MiB/s wr, 311 op/s Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.149 284030 INFO nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Creating config drive at /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.155 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): /usr/bin/mkisofs -o /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config -ldots -allow-lowercase -allow-multidot -l -publisher OpenStack Compute 27.5.2-0.20250829104910.6f8decf.el9 -quiet -J -r -V config-2 /tmp/tmp095z7url execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.286 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "/usr/bin/mkisofs -o /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config -ldots -allow-lowercase -allow-multidot -l -publisher OpenStack Compute 27.5.2-0.20250829104910.6f8decf.el9 -quiet -J -r -V config-2 /tmp/tmp095z7url" returned: 0 in 0.130s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.329 284030 DEBUG nova.storage.rbd_utils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] rbd image 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk.config does not exist __init__ /usr/lib/python3.9/site-packages/nova/storage/rbd_utils.py:80#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.334 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Running cmd (subprocess): rbd import --pool vms /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk.config --image-format=2 --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.559 284030 DEBUG nova.network.neutron [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Updating instance_info_cache with network_info: [] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.571 284030 DEBUG oslo_concurrency.processutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] CMD "rbd import --pool vms /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config 765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk.config --image-format=2 --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.237s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.571 284030 INFO nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Deleting local config drive /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677/disk.config because it was imported into RBD.#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.581 284030 INFO nova.compute.manager [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Took 1.72 seconds to deallocate network for instance.#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.619 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.620 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.623 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.003s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:11 localhost kernel: device tap20252d49-04 entered promiscuous mode Nov 27 05:04:11 localhost systemd-udevd[312529]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:04:11 localhost NetworkManager[5971]: [1764237851.6283] manager: (tap20252d49-04): new Tun device (/org/freedesktop/NetworkManager/Devices/24) Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.671 284030 INFO nova.scheduler.client.report [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Deleted allocations for instance ea7c65ed-7232-403d-84e6-2ead60aade7c#033[00m Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00104|binding|INFO|Claiming lport 20252d49-04fc-4922-8fc1-99b7e3b77a67 for this chassis. Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00105|binding|INFO|20252d49-04fc-4922-8fc1-99b7e3b77a67: Claiming fa:16:3e:75:1f:c2 10.100.0.8 Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00106|binding|INFO|Claiming lport 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 for this chassis. Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00107|binding|INFO|2161daa9-fc37-4b10-97d3-5c8dd774b7d3: Claiming fa:16:3e:69:7c:e1 19.80.0.52 Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.681 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:11 localhost NetworkManager[5971]: [1764237851.6856] device (tap20252d49-04): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 27 05:04:11 localhost NetworkManager[5971]: [1764237851.6863] device (tap20252d49-04): state change: unavailable -> disconnected (reason 'none', sys-iface-state: 'external') Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.696 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:69:7c:e1 19.80.0.52'], port_security=['fa:16:3e:69:7c:e1 19.80.0.52'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=['20252d49-04fc-4922-8fc1-99b7e3b77a67'], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-subport-463929020', 'neutron:cidrs': '19.80.0.52/24', 'neutron:device_id': '', 'neutron:device_owner': 'trunk:subport', 'neutron:mtu': '', 'neutron:network_name': 'neutron-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-subport-463929020', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '3', 'neutron:security_group_ids': '6d11bb1f-93da-41d7-8007-c66992d17e28', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[42], additional_encap=[], encap=[], mirror_rules=[], datapath=d463cc2b-c33c-4d8e-8d92-b80cf5ac63e7, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=2161daa9-fc37-4b10-97d3-5c8dd774b7d3) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.699 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:75:1f:c2 10.100.0.8'], port_security=['fa:16:3e:75:1f:c2 10.100.0.8'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-parent-2036430678', 'neutron:cidrs': '10.100.0.8/28', 'neutron:device_id': '765bb3f2-7b81-4d43-8eea-4f8b0e498677', 'neutron:device_owner': 'compute:nova', 'neutron:mtu': '', 'neutron:network_name': 'neutron-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-parent-2036430678', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '2', 'neutron:security_group_ids': '6d11bb1f-93da-41d7-8007-c66992d17e28', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c43ca390-01ce-4722-b559-7a1baab9099d, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=20252d49-04fc-4922-8fc1-99b7e3b77a67) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.700 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 in datapath 14034b34-9875-46a0-b548-1a34b924b9b9 bound to our chassis#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.704 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 8429fb01-77a6-4ffd-ac06-c71e08b7dcf0 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.704 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network 14034b34-9875-46a0-b548-1a34b924b9b9#033[00m Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00108|binding|INFO|Setting lport 20252d49-04fc-4922-8fc1-99b7e3b77a67 ovn-installed in OVS Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00109|binding|INFO|Setting lport 20252d49-04fc-4922-8fc1-99b7e3b77a67 up in Southbound Nov 27 05:04:11 localhost ovn_controller[156436]: 2025-11-27T10:04:11Z|00110|binding|INFO|Setting lport 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 up in Southbound Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.708 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:11 localhost systemd-machined[86026]: New machine qemu-4-instance-00000008. Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.715 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3bd6bfb7-8027-4e69-92c2-732216789b42]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.716 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Creating VETH tap14034b34-91 in ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9 namespace provision_datapath /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:665#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.719 162196 DEBUG neutron.privileged.agent.linux.ip_lib [-] Interface tap14034b34-90 not found in namespace None get_link_id /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:204#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.719 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[19f36cf1-8cbe-4639-856d-9e71058c784f]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.721 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0f546a00-ef43-4913-92dc-710cef820a3a]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost systemd[1]: Started Virtual Machine qemu-4-instance-00000008. Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.747 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[44dbd071-c152-41fc-b917-cad1b5a30ace]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.761 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b4c12e1a-9de5-454f-8068-318c2c1d02fa]: (4, ('net.ipv4.conf.all.promote_secondaries = 1\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.789 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[6693974d-5048-4452-9813-a74b3cb1cd40]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.791 284030 DEBUG oslo_concurrency.lockutils [None req-16e40d47-e76b-419d-be79-ad126cd642d5 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Lock "ea7c65ed-7232-403d-84e6-2ead60aade7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.211s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:11 localhost NetworkManager[5971]: [1764237851.7993] manager: (tap14034b34-90): new Veth device (/org/freedesktop/NetworkManager/Devices/25) Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.798 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3e0cdaf3-b0e4-4334-9222-094bbced59c8]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.834 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[45a17d46-c9c5-4a3f-93fa-3210f08f0cb4]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.838 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[5f0f1c5a-65ac-4dd7-83b5-4d5d684ef15c]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap14034b34-91: link becomes ready Nov 27 05:04:11 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap14034b34-90: link becomes ready Nov 27 05:04:11 localhost NetworkManager[5971]: [1764237851.8659] device (tap14034b34-90): carrier: link connected Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.873 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[06d14f16-6d9c-4b36-a86e-4065eaca26e5]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.890 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8db2db70-4953-4202-be9e-edd1226c21bc]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap14034b34-91'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:8c:75:e8'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 27], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208198, 'reachable_time': 24575, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 312963, 'error': None, 'target': 'ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.905 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8a64a38f-bc98-47cb-be4d-0f4f8f9cce5d]: (4, ({'family': 10, 'prefixlen': 64, 'flags': 192, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:fe8c:75e8'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 1208198, 'tstamp': 1208198}], ['IFA_FLAGS', 192]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 312971, 'error': None, 'target': 'ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'},)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.921 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[41e2ccf9-2f15-4683-bc26-1c888ca58a08]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap14034b34-91'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:8c:75:e8'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 27], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208198, 'reachable_time': 24575, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 0, 'sequence_number': 255, 'pid': 312980, 'error': None, 'target': 'ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.952 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[78834b75-88af-433d-83a4-eab1c047e639]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.991 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3b7eff86-e4d8-4c57-81c2-2f487488cec5]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.993 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap14034b34-90, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.993 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:04:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:11.993 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap14034b34-90, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.996 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:11 localhost kernel: device tap14034b34-90 entered promiscuous mode Nov 27 05:04:11 localhost nova_compute[284026]: 2025-11-27 10:04:11.999 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.000 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tap14034b34-90, col_values=(('external_ids', {'iface-id': '0b3b33c8-3d8e-4978-ae70-eedba54823eb'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.001 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:12 localhost ovn_controller[156436]: 2025-11-27T10:04:12Z|00111|binding|INFO|Releasing lport 0b3b33c8-3d8e-4978-ae70-eedba54823eb from this chassis (sb_readonly=0) Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.004 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/14034b34-9875-46a0-b548-1a34b924b9b9.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/14034b34-9875-46a0-b548-1a34b924b9b9.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.004 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.005 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[9cb3e081-b2f9-4777-a65e-812bc8a94f34]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.006 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: global Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-14034b34-9875-46a0-b548-1a34b924b9b9 Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: user root Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: group root Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/14034b34-9875-46a0-b548-1a34b924b9b9.pid.haproxy Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: daemon Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: defaults Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: log global Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: mode http Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: option httplog Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: listen listener Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID 14034b34-9875-46a0-b548-1a34b924b9b9 Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.007 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9', 'env', 'PROCESS_TAG=haproxy-14034b34-9875-46a0-b548-1a34b924b9b9', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/14034b34-9875-46a0-b548-1a34b924b9b9.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.023 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.099 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Started> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.100 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] VM Started (Lifecycle Event)#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.130 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.136 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Paused> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.136 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] VM Paused (Lifecycle Event)#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.162 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.166 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Synchronizing instance power state after lifecycle event "Paused"; current vm_state: building, current task_state: spawning, current DB power_state: 0, VM power_state: 3 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.194 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] During sync_power_state the instance has a pending task (spawning). Skip.#033[00m Nov 27 05:04:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e96 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:12 localhost podman[313040]: Nov 27 05:04:12 localhost podman[313040]: 2025-11-27 10:04:12.464172138 +0000 UTC m=+0.093550206 container create 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 05:04:12 localhost systemd[1]: Started libpod-conmon-47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed.scope. Nov 27 05:04:12 localhost podman[313040]: 2025-11-27 10:04:12.420504415 +0000 UTC m=+0.049882533 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 05:04:12 localhost systemd[1]: tmp-crun.coOZK0.mount: Deactivated successfully. Nov 27 05:04:12 localhost systemd[1]: Started libcrun container. Nov 27 05:04:12 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c5442b275610224ce17c224f1fe6193adc7d8b27a44666135d5ae38f8bef3e8b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:04:12 localhost podman[313040]: 2025-11-27 10:04:12.548245675 +0000 UTC m=+0.177623753 container init 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:04:12 localhost podman[313040]: 2025-11-27 10:04:12.556860539 +0000 UTC m=+0.186238607 container start 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:04:12 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [NOTICE] (313058) : New worker (313060) forked Nov 27 05:04:12 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [NOTICE] (313058) : Loading success. Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.601 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 20252d49-04fc-4922-8fc1-99b7e3b77a67 in datapath 2247f41d-32cc-48a7-83ff-88f84e00e09c unbound from our chassis#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.605 162092 INFO neutron.agent.ovn.metadata.agent [-] Provisioning metadata for network 2247f41d-32cc-48a7-83ff-88f84e00e09c#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.611 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[55b2641e-5455-4999-ba3b-7713685e314d]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.612 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Creating VETH tap2247f41d-31 in ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c namespace provision_datapath /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:665#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.615 162196 DEBUG neutron.privileged.agent.linux.ip_lib [-] Interface tap2247f41d-30 not found in namespace None get_link_id /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:204#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.616 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[43344cd5-ebd1-458f-be46-36ca1ac1e42e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.617 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[075ee26a-b9bf-49d0-ac07-590b8aaf36d6]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.691 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[22c04985-e9d5-40de-ad3e-54e25d15b9d3]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.700 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[38bee8d7-fa62-40f9-ae37-8487e02b06f1]: (4, ('net.ipv4.conf.all.promote_secondaries = 1\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.728 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[94e75507-e2af-4872-8219-81b0878c9bca]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.734 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[47394263-200c-41ce-8348-5ddab07f9a20]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost systemd-udevd[312948]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:04:12 localhost NetworkManager[5971]: [1764237852.7382] manager: (tap2247f41d-30): new Veth device (/org/freedesktop/NetworkManager/Devices/26) Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.763 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[d4a7060a-1556-4bed-b1ac-fe82c812a33c]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.771 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[6b523753-696f-4e0e-9564-2a690b0884a0]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost kernel: IPv6: ADDRCONF(NETDEV_CHANGE): tap2247f41d-30: link becomes ready Nov 27 05:04:12 localhost NetworkManager[5971]: [1764237852.7948] device (tap2247f41d-30): carrier: link connected Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.798 162207 DEBUG oslo.privsep.daemon [-] privsep: reply[7de82746-ca08-4f8a-beb8-14bfb604b0eb]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.815 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[374e8f09-470f-4419-8706-1c106a0ce535]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap2247f41d-31'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:48:6f:0e'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 28], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208291, 'reachable_time': 43030, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 313079, 'error': None, 'target': 'ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.832 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[cb3d1415-bcd6-4b8a-91e2-3bd230df5393]: (4, ({'family': 10, 'prefixlen': 64, 'flags': 192, 'scope': 253, 'index': 2, 'attrs': [['IFA_ADDRESS', 'fe80::f816:3eff:fe48:6f0e'], ['IFA_CACHEINFO', {'ifa_preferred': 4294967295, 'ifa_valid': 4294967295, 'cstamp': 1208291, 'tstamp': 1208291}], ['IFA_FLAGS', 192]], 'header': {'length': 72, 'type': 20, 'flags': 2, 'sequence_number': 255, 'pid': 313080, 'error': None, 'target': 'ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'stats': (0, 0, 0)}, 'event': 'RTM_NEWADDR'},)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.850 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[2c40818c-6ddd-45c5-9530-05f90ab7e271]: (4, [{'family': 0, '__align': (), 'ifi_type': 1, 'index': 2, 'flags': 69699, 'change': 0, 'attrs': [['IFLA_IFNAME', 'tap2247f41d-31'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UP'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 1500], ['IFLA_MIN_MTU', 68], ['IFLA_MAX_MTU', 65535], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 8], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 8], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 2], ['IFLA_CARRIER_UP_COUNT', 1], ['IFLA_CARRIER_DOWN_COUNT', 1], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', 'fa:16:3e:48:6f:0e'], ['IFLA_BROADCAST', 'ff:ff:ff:ff:ff:ff'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 90, 'tx_bytes': 90, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_LINKINFO', {'attrs': [['IFLA_INFO_KIND', 'veth']]}], ['IFLA_LINK_NETNSID', 0], ['IFLA_LINK', 28], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 0, 'nopolicy': 0, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208291, 'reachable_time': 43030, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 1500, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 0, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 1, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 1, 'inoctets': 76, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 1, 'outoctets': 76, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 1, 'outmcastpkts': 1, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 76, 'outmcastoctets': 76, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 1, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 1, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1400, 'type': 16, 'flags': 0, 'sequence_number': 255, 'pid': 313081, 'error': None, 'target': 'ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.882 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[176dcd49-348e-4390-8d19-471d5df3db74]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.954 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d23cd2b2-9590-4183-a002-e13673e76e4d]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.956 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap2247f41d-30, bridge=br-ex, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.957 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129#033[00m Nov 27 05:04:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:12.957 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): AddPortCommand(_result=None, bridge=br-int, port=tap2247f41d-30, may_exist=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:12 localhost nova_compute[284026]: 2025-11-27 10:04:12.997 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:13 localhost kernel: device tap2247f41d-30 entered promiscuous mode Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.004 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:13.005 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Interface, record=tap2247f41d-30, col_values=(('external_ids', {'iface-id': '5e75e77f-63bf-400e-98cb-3e3611a5e260'}),)) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.007 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:13 localhost ovn_controller[156436]: 2025-11-27T10:04:13Z|00112|binding|INFO|Releasing lport 5e75e77f-63bf-400e-98cb-3e3611a5e260 from this chassis (sb_readonly=0) Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.016 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.019 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:13.020 162092 DEBUG neutron.agent.linux.utils [-] Unable to access /var/lib/neutron/external/pids/2247f41d-32cc-48a7-83ff-88f84e00e09c.pid.haproxy; Error: [Errno 2] No such file or directory: '/var/lib/neutron/external/pids/2247f41d-32cc-48a7-83ff-88f84e00e09c.pid.haproxy' get_value_from_file /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:252#033[00m Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:13.021 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[56b9f63e-1d71-4d49-bf7a-77429d6abf80]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:13.022 162092 DEBUG neutron.agent.ovn.metadata.driver [-] haproxy_cfg = Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: global Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: log /dev/log local0 debug Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: log-tag haproxy-metadata-proxy-2247f41d-32cc-48a7-83ff-88f84e00e09c Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: user root Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: group root Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: maxconn 1024 Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: pidfile /var/lib/neutron/external/pids/2247f41d-32cc-48a7-83ff-88f84e00e09c.pid.haproxy Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: daemon Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: defaults Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: log global Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: mode http Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: option httplog Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: option dontlognull Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: option http-server-close Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: option forwardfor Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: retries 3 Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: timeout http-request 30s Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: timeout connect 30s Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: timeout client 32s Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: timeout server 32s Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: timeout http-keep-alive 30s Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: listen listener Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: bind 169.254.169.254:80 Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: server metadata /var/lib/neutron/metadata_proxy Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: http-request add-header X-OVN-Network-ID 2247f41d-32cc-48a7-83ff-88f84e00e09c Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: create_config_file /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/driver.py:107#033[00m Nov 27 05:04:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:13.023 162092 DEBUG neutron.agent.linux.utils [-] Running command: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'ip', 'netns', 'exec', 'ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'env', 'PROCESS_TAG=haproxy-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'haproxy', '-f', '/var/lib/neutron/ovn-metadata-proxy/2247f41d-32cc-48a7-83ff-88f84e00e09c.conf'] create_process /usr/lib/python3.9/site-packages/neutron/agent/linux/utils.py:84#033[00m Nov 27 05:04:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v145: 177 pgs: 177 active+clean; 271 MiB data, 970 MiB used, 41 GiB / 42 GiB avail; 3.4 MiB/s rd, 3.3 MiB/s wr, 263 op/s Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.136 284030 DEBUG nova.compute.manager [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.137 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.137 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.138 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.138 284030 DEBUG nova.compute.manager [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Processing event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 _process_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:10808#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.139 284030 DEBUG nova.compute.manager [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.139 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.142 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.143 284030 DEBUG oslo_concurrency.lockutils [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.149 284030 DEBUG nova.compute.manager [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] No waiting events found dispatching network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.150 284030 WARNING nova.compute.manager [req-99396a35-e380-4c75-a1ff-a57ef39d8e13 req-e373e881-768e-43a6-84d0-44aa70065823 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received unexpected event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 for instance with vm_state building and task_state spawning.#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.153 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Instance event wait completed in 1 seconds for network-vif-plugged wait_for_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:577#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.160 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Guest created on hypervisor spawn /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:4417#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.161 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Resumed> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.161 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] VM Resumed (Lifecycle Event)#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.168 284030 INFO nova.virt.libvirt.driver [-] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Instance spawned successfully.#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.168 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Attempting to register defaults for the following image properties: ['hw_cdrom_bus', 'hw_disk_bus', 'hw_input_bus', 'hw_pointer_model', 'hw_video_model', 'hw_vif_model'] _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:917#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.203 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.215 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Synchronizing instance power state after lifecycle event "Resumed"; current vm_state: building, current task_state: spawning, current DB power_state: 0, VM power_state: 1 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.226 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_cdrom_bus of sata _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.227 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_disk_bus of virtio _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.228 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_input_bus of usb _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.228 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_pointer_model of usbtablet _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.229 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_video_model of virtio _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.230 284030 DEBUG nova.virt.libvirt.driver [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Found default for hw_vif_model of virtio _register_undefined_instance_details /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:946#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.247 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] During sync_power_state the instance has a pending task (spawning). Skip.#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.303 284030 INFO nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Took 5.90 seconds to spawn the instance on the hypervisor.#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.304 284030 DEBUG nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.379 284030 INFO nova.compute.manager [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Took 7.00 seconds to build instance.#033[00m Nov 27 05:04:13 localhost nova_compute[284026]: 2025-11-27 10:04:13.405 284030 DEBUG oslo_concurrency.lockutils [None req-ee92a74f-474e-4a88-848d-d0b3cd3b1de9 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 7.110s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:13 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:13.438 2 INFO neutron.agent.securitygroups_rpc [None req-0308416e-44f7-4be9-afd6-527b62831db7 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Security group member updated ['d5a4dbad-2a9b-45ce-81ca-0430479e3c01']#033[00m Nov 27 05:04:13 localhost podman[313114]: Nov 27 05:04:13 localhost podman[313114]: 2025-11-27 10:04:13.54482843 +0000 UTC m=+0.096769112 container create 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:04:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:04:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:04:13 localhost systemd[1]: Started libpod-conmon-94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e.scope. Nov 27 05:04:13 localhost podman[313114]: 2025-11-27 10:04:13.49607065 +0000 UTC m=+0.048011362 image pull quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified Nov 27 05:04:13 localhost systemd[1]: Started libcrun container. Nov 27 05:04:13 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/990de230d5bdaa49bdaeacb82490cbc58663eea506c2e803528dd6b6d50f18cb/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:04:13 localhost podman[313114]: 2025-11-27 10:04:13.63524841 +0000 UTC m=+0.187189102 container init 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:13 localhost podman[313114]: 2025-11-27 10:04:13.649516076 +0000 UTC m=+0.201456768 container start 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:04:13 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [NOTICE] (313153) : New worker (313155) forked Nov 27 05:04:13 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [NOTICE] (313153) : Loading success. Nov 27 05:04:13 localhost podman[313129]: 2025-11-27 10:04:13.736978025 +0000 UTC m=+0.145145663 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:04:13 localhost podman[313129]: 2025-11-27 10:04:13.752069094 +0000 UTC m=+0.160236762 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:04:13 localhost podman[313128]: 2025-11-27 10:04:13.708598957 +0000 UTC m=+0.123721883 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, container_name=ovn_controller, io.buildah.version=1.41.3, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:13 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:04:13 localhost podman[313128]: 2025-11-27 10:04:13.788293005 +0000 UTC m=+0.203415971 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller) Nov 27 05:04:13 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v146: 177 pgs: 177 active+clean; 271 MiB data, 970 MiB used, 41 GiB / 42 GiB avail; 2.7 MiB/s rd, 2.7 MiB/s wr, 209 op/s Nov 27 05:04:15 localhost nova_compute[284026]: 2025-11-27 10:04:15.524 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:15.784 2 INFO neutron.agent.securitygroups_rpc [None req-abf2a09d-90e4-4c13-b9e6-bde9b18578f3 730775ba58804b5b8977106a8a8aa97c 0780cc2b55f74fbd8c97a1f5f39cd8cc - - default default] Security group member updated ['d5a4dbad-2a9b-45ce-81ca-0430479e3c01']#033[00m Nov 27 05:04:15 localhost nova_compute[284026]: 2025-11-27 10:04:15.845 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:16 localhost nova_compute[284026]: 2025-11-27 10:04:16.491 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Check if temp file /var/lib/nova/instances/tmp8uqckvp7 exists to indicate shared storage is being used for migration. Exists? False _check_shared_storage_test_file /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10065#033[00m Nov 27 05:04:16 localhost nova_compute[284026]: 2025-11-27 10:04:16.492 284030 DEBUG nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] source check data is LibvirtLiveMigrateData(bdms=,block_migration=False,disk_available_mb=12288,disk_over_commit=False,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp8uqckvp7',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='765bb3f2-7b81-4d43-8eea-4f8b0e498677',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=,old_vol_attachment_ids=,serial_listen_addr=None,serial_listen_ports=,src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=,target_connect_addr=,vifs=[VIFMigrateData],wait_for_vif_plugged=) check_can_live_migrate_source /usr/lib/python3.9/site-packages/nova/compute/manager.py:8587#033[00m Nov 27 05:04:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e97 e97: 6 total, 6 up, 6 in Nov 27 05:04:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v148: 177 pgs: 177 active+clean; 350 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 10 MiB/s rd, 8.5 MiB/s wr, 417 op/s Nov 27 05:04:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e97 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e98 e98: 6 total, 6 up, 6 in Nov 27 05:04:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:18.028 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:04:17Z, description=, device_id=694a3d3a-24c9-4116-8de6-00280384fa42, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=a5c761a3-f32e-44e7-a9b0-5879cdf355ef, ip_allocation=immediate, mac_address=fa:16:3e:c2:9f:00, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=713, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:04:17Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:04:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:18.078 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=11, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=10) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:18.080 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 9 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:04:18 localhost nova_compute[284026]: 2025-11-27 10:04:18.079 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:18 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 6 addresses Nov 27 05:04:18 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:18 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:18 localhost podman[313207]: 2025-11-27 10:04:18.30902745 +0000 UTC m=+0.077978453 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:04:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:18.578 265123 INFO neutron.agent.dhcp.agent [None req-2cab829f-17fd-435f-bcf9-43004582d378 - - - - - -] DHCP configuration for ports {'a5c761a3-f32e-44e7-a9b0-5879cdf355ef'} is completed#033[00m Nov 27 05:04:18 localhost nova_compute[284026]: 2025-11-27 10:04:18.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:18 localhost nova_compute[284026]: 2025-11-27 10:04:18.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:18 localhost nova_compute[284026]: 2025-11-27 10:04:18.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e99 e99: 6 total, 6 up, 6 in Nov 27 05:04:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v151: 177 pgs: 177 active+clean; 350 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 10 MiB/s rd, 7.8 MiB/s wr, 277 op/s Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.689 284030 DEBUG nova.compute.manager [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.690 284030 DEBUG oslo_concurrency.lockutils [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.691 284030 DEBUG oslo_concurrency.lockutils [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.691 284030 DEBUG oslo_concurrency.lockutils [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.691 284030 DEBUG nova.compute.manager [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] No waiting events found dispatching network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.692 284030 DEBUG nova.compute.manager [req-20897695-e37d-420e-8e77-d9c4194a8ff1 req-f6e9b2cb-2ff2-47e0-9fae-9c1eab711cee c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 for instance with task_state migrating. _process_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:10826#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:19 localhost nova_compute[284026]: 2025-11-27 10:04:19.763 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e100 e100: 6 total, 6 up, 6 in Nov 27 05:04:20 localhost nova_compute[284026]: 2025-11-27 10:04:20.559 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e101 e101: 6 total, 6 up, 6 in Nov 27 05:04:20 localhost nova_compute[284026]: 2025-11-27 10:04:20.848 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:21 localhost ovn_controller[156436]: 2025-11-27T10:04:21Z|00113|binding|INFO|Releasing lport 0b3b33c8-3d8e-4978-ae70-eedba54823eb from this chassis (sb_readonly=0) Nov 27 05:04:21 localhost ovn_controller[156436]: 2025-11-27T10:04:21Z|00114|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:21 localhost ovn_controller[156436]: 2025-11-27T10:04:21Z|00115|binding|INFO|Releasing lport 5e75e77f-63bf-400e-98cb-3e3611a5e260 from this chassis (sb_readonly=0) Nov 27 05:04:21 localhost systemd[1]: tmp-crun.wdTeRA.mount: Deactivated successfully. Nov 27 05:04:21 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:04:21 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:21 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:21 localhost podman[313244]: 2025-11-27 10:04:21.063632725 +0000 UTC m=+0.074318774 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:04:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v154: 177 pgs: 177 active+clean; 271 MiB data, 933 MiB used, 41 GiB / 42 GiB avail; 7.9 MiB/s rd, 24 KiB/s wr, 493 op/s Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.119 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.424 284030 INFO nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Took 3.93 seconds for pre_live_migration on destination host np0005537444.localdomain.#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.425 284030 DEBUG nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Instance event wait completed in 0 seconds for wait_for_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:577#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.460 284030 DEBUG nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] live_migration data is LibvirtLiveMigrateData(bdms=[],block_migration=False,disk_available_mb=12288,disk_over_commit=False,dst_numa_info=,dst_supports_numa_live_migration=,dst_wants_file_backed_memory=False,file_backed_memory_discard=,filename='tmp8uqckvp7',graphics_listen_addr_spice=127.0.0.1,graphics_listen_addr_vnc=::,image_type='rbd',instance_relative_path='765bb3f2-7b81-4d43-8eea-4f8b0e498677',is_shared_block_storage=True,is_shared_instance_path=False,is_volume_backed=False,migration=Migration(29b259ea-ef69-4cbd-be42-1ec4a8fedcae),old_vol_attachment_ids={},serial_listen_addr=None,serial_listen_ports=[],src_supports_native_luks=,src_supports_numa_live_migration=,supported_perf_events=[],target_connect_addr=None,vifs=[VIFMigrateData],wait_for_vif_plugged=True) _do_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:8939#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.465 284030 DEBUG nova.objects.instance [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lazy-loading 'migration_context' on Instance uuid 765bb3f2-7b81-4d43-8eea-4f8b0e498677 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.469 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Starting monitoring of live migration _live_migration /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10639#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.475 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Operation thread is still running _live_migration_monitor /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10440#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.476 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migration not running yet _live_migration_monitor /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10449#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.500 284030 DEBUG nova.virt.libvirt.vif [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='True',created_at=2025-11-27T10:04:05Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='tempest-LiveMigrationTest-server-1077612390',display_name='tempest-LiveMigrationTest-server-1077612390',ec2_ids=,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537446.localdomain',hostname='tempest-livemigrationtest-server-1077612390',id=8,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T10:04:13Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=None,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=InstancePCIRequests,power_state=1,progress=0,project_id='a6d84e90a5fa4f74afa4103832d90ac2',ramdisk_id='',reservation_id='r-mchq7ras',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='q35',image_hw_pointer_model='usbtablet',image_hw_rng_model='virtio',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',owner_project_name='tempest-LiveMigrationTest-1823599681',owner_user_name='tempest-LiveMigrationTest-1823599681-project-member'},tags=,task_state='migrating',terminated_at=None,trusted_certs=,updated_at=2025-11-27T10:04:13Z,user_data=None,user_id='edb03acaf09647f6b6b5465b16878a77',uuid=765bb3f2-7b81-4d43-8eea-4f8b0e498677,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system"}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} virt_type=kvm get_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:563#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.502 284030 DEBUG nova.network.os_vif_util [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Converting VIF {"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system"}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {"os_vif_delegation": true}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.506 284030 DEBUG nova.network.os_vif_util [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.507 284030 DEBUG nova.virt.libvirt.migration [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updating guest XML with vif config: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: Nov 27 05:04:21 localhost nova_compute[284026]: _update_vif_xml /usr/lib/python3.9/site-packages/nova/virt/libvirt/migration.py:388#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.509 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] About to invoke the migrate API _live_migration_operation /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10272#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.734 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.741 284030 DEBUG nova.compute.manager [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.741 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.742 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.742 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.743 284030 DEBUG nova.compute.manager [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] No waiting events found dispatching network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.743 284030 WARNING nova.compute.manager [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received unexpected event network-vif-plugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 for instance with vm_state active and task_state migrating.#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.744 284030 DEBUG nova.compute.manager [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-changed-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.744 284030 DEBUG nova.compute.manager [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Refreshing instance network info cache due to event network-changed-20252d49-04fc-4922-8fc1-99b7e3b77a67. external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11053#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.745 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.745 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquired lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.746 284030 DEBUG nova.network.neutron [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Refreshing network info cache for port 20252d49-04fc-4922-8fc1-99b7e3b77a67 _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2007#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.982 284030 DEBUG nova.virt.libvirt.migration [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Current None elapsed 0 steps [(0, 50), (150, 95), (300, 140), (450, 185), (600, 230), (750, 275), (900, 320), (1050, 365), (1200, 410), (1350, 455), (1500, 500)] update_downtime /usr/lib/python3.9/site-packages/nova/virt/libvirt/migration.py:512#033[00m Nov 27 05:04:21 localhost nova_compute[284026]: 2025-11-27 10:04:21.983 284030 INFO nova.virt.libvirt.migration [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Increasing downtime to 50 ms after 0 sec elapsed time#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.075 284030 INFO nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migration running for 0 secs, memory 100% remaining (bytes processed=0, remaining=0, total=0); disk 100% remaining (bytes processed=0, remaining=0, total=0).#033[00m Nov 27 05:04:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e101 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.652 284030 DEBUG nova.virt.libvirt.migration [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Current 50 elapsed 1 steps [(0, 50), (150, 95), (300, 140), (450, 185), (600, 230), (750, 275), (900, 320), (1050, 365), (1200, 410), (1350, 455), (1500, 500)] update_downtime /usr/lib/python3.9/site-packages/nova/virt/libvirt/migration.py:512#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.653 284030 DEBUG nova.virt.libvirt.migration [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Downtime does not need to change update_downtime /usr/lib/python3.9/site-packages/nova/virt/libvirt/migration.py:525#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.663 284030 DEBUG nova.network.neutron [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updated VIF entry in instance network info cache for port 20252d49-04fc-4922-8fc1-99b7e3b77a67. _build_network_info_model /usr/lib/python3.9/site-packages/nova/network/neutron.py:3482#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.664 284030 DEBUG nova.network.neutron [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updating instance_info_cache with network_info: [{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {"migrating_to": "np0005537444.localdomain"}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.682 284030 DEBUG oslo_concurrency.lockutils [req-358aa98b-b4e1-478b-8ed7-d5b6d82b8ada req-a9c2f48e-4d6a-4b53-8193-054d7562ced0 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Releasing lock "refresh_cache-765bb3f2-7b81-4d43-8eea-4f8b0e498677" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.729 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.917 284030 DEBUG nova.virt.driver [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] Emitting event Paused> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.918 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] VM Paused (Lifecycle Event)#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.946 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.951 284030 DEBUG nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Synchronizing instance power state after lifecycle event "Paused"; current vm_state: active, current task_state: migrating, current DB power_state: 1, VM power_state: 3 handle_lifecycle_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:1396#033[00m Nov 27 05:04:22 localhost nova_compute[284026]: 2025-11-27 10:04:22.985 284030 INFO nova.compute.manager [None req-2a71cb04-5d5d-4c1d-8d4c-bd6faa0cf1e6 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] During sync_power_state the instance has a pending task (migrating). Skip.#033[00m Nov 27 05:04:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e102 e102: 6 total, 6 up, 6 in Nov 27 05:04:23 localhost kernel: device tap20252d49-04 left promiscuous mode Nov 27 05:04:23 localhost NetworkManager[5971]: [1764237863.0525] device (tap20252d49-04): state change: disconnected -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00116|binding|INFO|Releasing lport 20252d49-04fc-4922-8fc1-99b7e3b77a67 from this chassis (sb_readonly=0) Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00117|binding|INFO|Setting lport 20252d49-04fc-4922-8fc1-99b7e3b77a67 down in Southbound Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00118|binding|INFO|Releasing lport 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 from this chassis (sb_readonly=0) Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00119|binding|INFO|Setting lport 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 down in Southbound Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.106 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00120|binding|INFO|Removing iface tap20252d49-04 ovn-installed in OVS Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.109 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v156: 177 pgs: 177 active+clean; 271 MiB data, 933 MiB used, 41 GiB / 42 GiB avail; 7.2 MiB/s rd, 22 KiB/s wr, 453 op/s Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.116 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:69:7c:e1 19.80.0.52'], port_security=['fa:16:3e:69:7c:e1 19.80.0.52'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=['20252d49-04fc-4922-8fc1-99b7e3b77a67'], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-subport-463929020', 'neutron:cidrs': '19.80.0.52/24', 'neutron:device_id': '', 'neutron:device_owner': 'trunk:subport', 'neutron:mtu': '', 'neutron:network_name': 'neutron-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-subport-463929020', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '3', 'neutron:security_group_ids': '6d11bb1f-93da-41d7-8007-c66992d17e28', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[42], additional_encap=[], encap=[], mirror_rules=[], datapath=d463cc2b-c33c-4d8e-8d92-b80cf5ac63e7, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=2161daa9-fc37-4b10-97d3-5c8dd774b7d3) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00121|binding|INFO|Releasing lport 0b3b33c8-3d8e-4978-ae70-eedba54823eb from this chassis (sb_readonly=0) Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00122|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:23 localhost ovn_controller[156436]: 2025-11-27T10:04:23Z|00123|binding|INFO|Releasing lport 5e75e77f-63bf-400e-98cb-3e3611a5e260 from this chassis (sb_readonly=0) Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.120 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:75:1f:c2 10.100.0.8'], port_security=['fa:16:3e:75:1f:c2 10.100.0.8'], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain,np0005537444.localdomain', 'activation-strategy': 'rarp', 'additional-chassis-activated': 'e11ccfbe-bbec-4f91-a55a-f7b4704203e7'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'name': 'tempest-parent-2036430678', 'neutron:cidrs': '10.100.0.8/28', 'neutron:device_id': '765bb3f2-7b81-4d43-8eea-4f8b0e498677', 'neutron:device_owner': 'compute:nova', 'neutron:mtu': '', 'neutron:network_name': 'neutron-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'neutron:port_capabilities': '', 'neutron:port_name': 'tempest-parent-2036430678', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '8', 'neutron:security_group_ids': '6d11bb1f-93da-41d7-8007-c66992d17e28', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c43ca390-01ce-4722-b559-7a1baab9099d, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=20252d49-04fc-4922-8fc1-99b7e3b77a67) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.123 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 2161daa9-fc37-4b10-97d3-5c8dd774b7d3 in datapath 14034b34-9875-46a0-b548-1a34b924b9b9 unbound from our chassis#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.126 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 8429fb01-77a6-4ffd-ac06-c71e08b7dcf0 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.126 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 14034b34-9875-46a0-b548-1a34b924b9b9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.127 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d013057a-2045-4ffe-a0d8-fbfff81dfd33]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.128 162092 INFO neutron.agent.ovn.metadata.agent [-] Cleaning up ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9 namespace which is not needed anymore#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.132 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost systemd[1]: machine-qemu\x2d4\x2dinstance\x2d00000008.scope: Deactivated successfully. Nov 27 05:04:23 localhost systemd[1]: machine-qemu\x2d4\x2dinstance\x2d00000008.scope: Consumed 10.273s CPU time. Nov 27 05:04:23 localhost systemd-machined[86026]: Machine qemu-4-instance-00000008 terminated. Nov 27 05:04:23 localhost journal[205316]: Unable to get XATTR trusted.libvirt.security.ref_selinux on vms/765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk: No such file or directory Nov 27 05:04:23 localhost journal[205316]: Unable to get XATTR trusted.libvirt.security.ref_dac on vms/765bb3f2-7b81-4d43-8eea-4f8b0e498677_disk: No such file or directory Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.215 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.222 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.235 284030 DEBUG nova.virt.libvirt.guest [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Domain has shutdown/gone away: Requested operation is not valid: domain is not running get_job_info /usr/lib/python3.9/site-packages/nova/virt/libvirt/guest.py:688#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.235 284030 INFO nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migration operation has completed#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.236 284030 INFO nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] _post_live_migration() is started..#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.249 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migrate API has completed _live_migration_operation /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10279#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.249 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migration operation thread has finished _live_migration_operation /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10327#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.249 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migration operation thread notification thread_finished /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10630#033[00m Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [NOTICE] (313058) : haproxy version is 2.8.14-c23fe91 Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [NOTICE] (313058) : path to executable is /usr/sbin/haproxy Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [WARNING] (313058) : Exiting Master process... Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [ALERT] (313058) : Current worker (313060) exited with code 143 (Terminated) Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9[313054]: [WARNING] (313058) : All workers exited. Exiting... (0) Nov 27 05:04:23 localhost systemd[1]: libpod-47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed.scope: Deactivated successfully. Nov 27 05:04:23 localhost podman[313303]: 2025-11-27 10:04:23.30950863 +0000 UTC m=+0.059836462 container died 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:23 localhost podman[313303]: 2025-11-27 10:04:23.349821871 +0000 UTC m=+0.100149663 container cleanup 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:23 localhost podman[313317]: 2025-11-27 10:04:23.381469929 +0000 UTC m=+0.064288852 container cleanup 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:04:23 localhost systemd[1]: libpod-conmon-47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed.scope: Deactivated successfully. Nov 27 05:04:23 localhost podman[313331]: 2025-11-27 10:04:23.440480548 +0000 UTC m=+0.071475808 container remove 47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.446 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e3ae935c-f11a-4ea6-aac8-b46d48745f78]: (4, ('Thu Nov 27 10:04:23 AM UTC 2025 Stopping container neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9 (47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed)\n47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed\nThu Nov 27 10:04:23 AM UTC 2025 Deleting container neutron-haproxy-ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9 (47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed)\n47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.449 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e5d0bfe5-9fbd-4d80-a6a0-87c6fc65329e]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.450 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap14034b34-90, bridge=None, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.453 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost kernel: device tap14034b34-90 left promiscuous mode Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.471 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.479 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e6c9f95e-0d7e-40c4-85b6-44f87b019348]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.495 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ccfab03c-bb5c-4451-ad3f-8d2aff290078]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.496 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[bae222f9-63c8-4e7c-adf8-f0d9f2602b9e]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.506 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3910b5bf-8e5e-41b0-9d93-875a22284608]: (4, [{'family': 0, '__align': (), 'ifi_type': 772, 'index': 1, 'flags': 65609, 'change': 0, 'attrs': [['IFLA_IFNAME', 'lo'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UNKNOWN'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 65536], ['IFLA_MIN_MTU', 0], ['IFLA_MAX_MTU', 0], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 1], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 1], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 0], ['IFLA_CARRIER_UP_COUNT', 0], ['IFLA_CARRIER_DOWN_COUNT', 0], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', '00:00:00:00:00:00'], ['IFLA_BROADCAST', '00:00:00:00:00:00'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 1, 'nopolicy': 1, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208190, 'reachable_time': 15209, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 65536, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 4294967295, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 4294967295, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 0, 'inoctets': 0, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 0, 'outoctets': 0, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 0, 'outmcastpkts': 0, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 0, 'outmcastoctets': 0, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 0, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 0, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1356, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 313356, 'error': None, 'target': 'ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.509 162227 DEBUG neutron.privileged.agent.linux.ip_lib [-] Namespace ovnmeta-14034b34-9875-46a0-b548-1a34b924b9b9 deleted. remove_netns /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:607#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.509 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[05da40c7-bcfc-40b5-b83f-6df3a7efcff9]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.509 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 20252d49-04fc-4922-8fc1-99b7e3b77a67 in datapath 2247f41d-32cc-48a7-83ff-88f84e00e09c unbound from our chassis#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.511 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 2247f41d-32cc-48a7-83ff-88f84e00e09c, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.512 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[62979e87-e67b-421e-8bfe-b20cdfc3c6f2]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.512 162092 INFO neutron.agent.ovn.metadata.agent [-] Cleaning up ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c namespace which is not needed anymore#033[00m Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [NOTICE] (313153) : haproxy version is 2.8.14-c23fe91 Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [NOTICE] (313153) : path to executable is /usr/sbin/haproxy Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [WARNING] (313153) : Exiting Master process... Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [WARNING] (313153) : Exiting Master process... Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [ALERT] (313153) : Current worker (313155) exited with code 143 (Terminated) Nov 27 05:04:23 localhost neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c[313135]: [WARNING] (313153) : All workers exited. Exiting... (0) Nov 27 05:04:23 localhost systemd[1]: libpod-94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e.scope: Deactivated successfully. Nov 27 05:04:23 localhost podman[313375]: 2025-11-27 10:04:23.697322434 +0000 UTC m=+0.082029023 container died 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:04:23 localhost podman[313375]: 2025-11-27 10:04:23.740629578 +0000 UTC m=+0.125336187 container cleanup 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:04:23 localhost podman[313388]: 2025-11-27 10:04:23.766973161 +0000 UTC m=+0.063119750 container cleanup 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:04:23 localhost systemd[1]: libpod-conmon-94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e.scope: Deactivated successfully. Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.818 284030 DEBUG nova.virt.driver [-] Emitting event Stopped> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.818 284030 INFO nova.compute.manager [-] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] VM Stopped (Lifecycle Event)#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.843 284030 DEBUG nova.compute.manager [None req-bed1add8-b14a-4a91-83d7-45d0f10587aa - - - - - -] [instance: ea7c65ed-7232-403d-84e6-2ead60aade7c] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:23 localhost podman[313404]: 2025-11-27 10:04:23.846973138 +0000 UTC m=+0.084222902 container remove 94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.853 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d8bcaef2-8046-4639-a5f9-a53c72f272b0]: (4, ('Thu Nov 27 10:04:23 AM UTC 2025 Stopping container neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c (94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e)\n94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e\nThu Nov 27 10:04:23 AM UTC 2025 Deleting container neutron-haproxy-ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c (94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e)\n94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e\n', '', 0)) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.855 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[93432096-ece9-4528-aa7b-8a7a4e2ecf7f]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.857 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap2247f41d-30, bridge=None, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:23 localhost kernel: device tap2247f41d-30 left promiscuous mode Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.860 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost nova_compute[284026]: 2025-11-27 10:04:23.871 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.875 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[1492f69a-d79f-44a6-8e2c-0d22de397dc9]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.899 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[31d28b1b-2774-4217-8b33-a1be02bfce0f]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.901 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d17cba65-e209-464f-82cd-fd6eefc7c31f]: (4, True) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.919 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0e773701-6fb7-45dc-9e2a-e9da050207fd]: (4, [{'family': 0, '__align': (), 'ifi_type': 772, 'index': 1, 'flags': 65609, 'change': 0, 'attrs': [['IFLA_IFNAME', 'lo'], ['IFLA_TXQLEN', 1000], ['IFLA_OPERSTATE', 'UNKNOWN'], ['IFLA_LINKMODE', 0], ['IFLA_MTU', 65536], ['IFLA_MIN_MTU', 0], ['IFLA_MAX_MTU', 0], ['IFLA_GROUP', 0], ['IFLA_PROMISCUITY', 0], ['IFLA_NUM_TX_QUEUES', 1], ['IFLA_GSO_MAX_SEGS', 65535], ['IFLA_GSO_MAX_SIZE', 65536], ['IFLA_GRO_MAX_SIZE', 65536], ['IFLA_TSO_MAX_SIZE', 524280], ['IFLA_TSO_MAX_SEGS', 65535], ['IFLA_NUM_RX_QUEUES', 1], ['IFLA_CARRIER', 1], ['IFLA_QDISC', 'noqueue'], ['IFLA_CARRIER_CHANGES', 0], ['IFLA_CARRIER_UP_COUNT', 0], ['IFLA_CARRIER_DOWN_COUNT', 0], ['IFLA_PROTO_DOWN', 0], ['IFLA_MAP', {'mem_start': 0, 'mem_end': 0, 'base_addr': 0, 'irq': 0, 'dma': 0, 'port': 0}], ['IFLA_ADDRESS', '00:00:00:00:00:00'], ['IFLA_BROADCAST', '00:00:00:00:00:00'], ['IFLA_STATS64', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_STATS', {'rx_packets': 1, 'tx_packets': 1, 'rx_bytes': 28, 'tx_bytes': 28, 'rx_errors': 0, 'tx_errors': 0, 'rx_dropped': 0, 'tx_dropped': 0, 'multicast': 0, 'collisions': 0, 'rx_length_errors': 0, 'rx_over_errors': 0, 'rx_crc_errors': 0, 'rx_frame_errors': 0, 'rx_fifo_errors': 0, 'rx_missed_errors': 0, 'tx_aborted_errors': 0, 'tx_carrier_errors': 0, 'tx_fifo_errors': 0, 'tx_heartbeat_errors': 0, 'tx_window_errors': 0, 'rx_compressed': 0, 'tx_compressed': 0}], ['IFLA_XDP', {'attrs': [['IFLA_XDP_ATTACHED', None]]}], ['IFLA_AF_SPEC', {'attrs': [['AF_INET', {'dummy': 65668, 'forwarding': 1, 'mc_forwarding': 0, 'proxy_arp': 0, 'accept_redirects': 0, 'secure_redirects': 0, 'send_redirects': 0, 'shared_media': 1, 'rp_filter': 1, 'accept_source_route': 0, 'bootp_relay': 0, 'log_martians': 1, 'tag': 0, 'arpfilter': 0, 'medium_id': 0, 'noxfrm': 1, 'nopolicy': 1, 'force_igmp_version': 0, 'arp_announce': 0, 'arp_ignore': 0, 'promote_secondaries': 1, 'arp_accept': 0, 'arp_notify': 0, 'accept_local': 0, 'src_vmark': 0, 'proxy_arp_pvlan': 0, 'route_localnet': 0, 'igmpv2_unsolicited_report_interval': 10000, 'igmpv3_unsolicited_report_interval': 1000}], ['AF_INET6', {'attrs': [['IFLA_INET6_FLAGS', 2147483648], ['IFLA_INET6_CACHEINFO', {'max_reasm_len': 65535, 'tstamp': 1208283, 'reachable_time': 30895, 'retrans_time': 1000}], ['IFLA_INET6_CONF', {'forwarding': 0, 'hop_limit': 64, 'mtu': 65536, 'accept_ra': 1, 'accept_redirects': 1, 'autoconf': 1, 'dad_transmits': 1, 'router_solicitations': 4294967295, 'router_solicitation_interval': 4000, 'router_solicitation_delay': 1000, 'use_tempaddr': 4294967295, 'temp_valid_lft': 604800, 'temp_preferred_lft': 86400, 'regen_max_retry': 3, 'max_desync_factor': 600, 'max_addresses': 16, 'force_mld_version': 0, 'accept_ra_defrtr': 1, 'accept_ra_pinfo': 1, 'accept_ra_rtr_pref': 1, 'router_probe_interval': 60000, 'accept_ra_rt_info_max_plen': 0, 'proxy_ndp': 0, 'optimistic_dad': 0, 'accept_source_route': 0, 'mc_forwarding': 0, 'disable_ipv6': 0, 'accept_dad': 4294967295, 'force_tllao': 0, 'ndisc_notify': 0}], ['IFLA_INET6_STATS', {'num': 37, 'inpkts': 0, 'inoctets': 0, 'indelivers': 0, 'outforwdatagrams': 0, 'outpkts': 0, 'outoctets': 0, 'inhdrerrors': 0, 'intoobigerrors': 0, 'innoroutes': 0, 'inaddrerrors': 0, 'inunknownprotos': 0, 'intruncatedpkts': 0, 'indiscards': 0, 'outdiscards': 0, 'outnoroutes': 0, 'reasmtimeout': 0, 'reasmreqds': 0, 'reasmoks': 0, 'reasmfails': 0, 'fragoks': 0, 'fragfails': 0, 'fragcreates': 0, 'inmcastpkts': 0, 'outmcastpkts': 0, 'inbcastpkts': 0, 'outbcastpkts': 0, 'inmcastoctets': 0, 'outmcastoctets': 0, 'inbcastoctets': 0, 'outbcastoctets': 0, 'csumerrors': 0, 'noectpkts': 0, 'ect1pkts': 0, 'ect0pkts': 0, 'cepkts': 0}], ['IFLA_INET6_ICMP6STATS', {'num': 7, 'inmsgs': 0, 'inerrors': 0, 'outmsgs': 0, 'outerrors': 0, 'csumerrors': 0}], ['IFLA_INET6_TOKEN', '::'], ['IFLA_INET6_ADDR_GEN_MODE', 0]]}]]}]], 'header': {'length': 1356, 'type': 16, 'flags': 2, 'sequence_number': 255, 'pid': 313426, 'error': None, 'target': 'ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c', 'stats': (0, 0, 0)}, 'state': 'up', 'event': 'RTM_NEWLINK'}]) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.921 162227 DEBUG neutron.privileged.agent.linux.ip_lib [-] Namespace ovnmeta-2247f41d-32cc-48a7-83ff-88f84e00e09c deleted. remove_netns /usr/lib/python3.9/site-packages/neutron/privileged/agent/linux/ip_lib.py:607#033[00m Nov 27 05:04:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:23.921 162227 DEBUG oslo.privsep.daemon [-] privsep: reply[3c906fdf-0362-4e96-849e-5b8d1b187669]: (4, None) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.252 284030 DEBUG nova.compute.manager [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 external_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:11048#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.253 284030 DEBUG oslo_concurrency.lockutils [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.253 284030 DEBUG oslo_concurrency.lockutils [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.254 284030 DEBUG oslo_concurrency.lockutils [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.254 284030 DEBUG nova.compute.manager [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] No waiting events found dispatching network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 pop_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:320#033[00m Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.254 284030 DEBUG nova.compute.manager [req-38750e96-3c2a-47eb-9ac1-0739b0e14185 req-d5ea841e-dcc0-4fb6-8ed9-91ac51a964c9 c155274c02da4a729a1e114d4e1317b8 ece81c3214184c54956c375af05c2f4f - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Received event network-vif-unplugged-20252d49-04fc-4922-8fc1-99b7e3b77a67 for instance with task_state migrating. _process_instance_event /usr/lib/python3.9/site-packages/nova/compute/manager.py:10826#033[00m Nov 27 05:04:24 localhost systemd[1]: var-lib-containers-storage-overlay-990de230d5bdaa49bdaeacb82490cbc58663eea506c2e803528dd6b6d50f18cb-merged.mount: Deactivated successfully. Nov 27 05:04:24 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-94922388f4767565f06284f967916ac99b214cb2c1ef3d2fb31499a1f2fb6f3e-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:24 localhost systemd[1]: run-netns-ovnmeta\x2d2247f41d\x2d32cc\x2d48a7\x2d83ff\x2d88f84e00e09c.mount: Deactivated successfully. Nov 27 05:04:24 localhost systemd[1]: var-lib-containers-storage-overlay-c5442b275610224ce17c224f1fe6193adc7d8b27a44666135d5ae38f8bef3e8b-merged.mount: Deactivated successfully. Nov 27 05:04:24 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-47402fa509ca5fd02743d30361be28196244f249bfa655286c25007e8fbc55ed-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:24 localhost systemd[1]: run-netns-ovnmeta\x2d14034b34\x2d9875\x2d46a0\x2db548\x2d1a34b924b9b9.mount: Deactivated successfully. Nov 27 05:04:24 localhost nova_compute[284026]: 2025-11-27 10:04:24.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e103 e103: 6 total, 6 up, 6 in Nov 27 05:04:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v158: 177 pgs: 177 active+clean; 271 MiB data, 933 MiB used, 41 GiB / 42 GiB avail; 5.9 MiB/s rd, 18 KiB/s wr, 370 op/s Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.380 284030 DEBUG nova.network.neutron [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Activated binding for port 20252d49-04fc-4922-8fc1-99b7e3b77a67 and host np0005537444.localdomain migrate_instance_start /usr/lib/python3.9/site-packages/nova/network/neutron.py:3181#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.381 284030 DEBUG nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Calling driver.post_live_migration_at_source with original source VIFs from migrate_data: [{"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] _post_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:9326#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.383 284030 DEBUG nova.virt.libvirt.vif [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] vif_type=ovs instance=Instance(access_ip_v4=None,access_ip_v6=None,architecture=None,auto_disk_config=False,availability_zone='nova',cell_name=None,cleaned=False,config_drive='True',created_at=2025-11-27T10:04:05Z,default_ephemeral_device=None,default_swap_device=None,deleted=False,deleted_at=None,device_metadata=,disable_terminate=False,display_description='tempest-LiveMigrationTest-server-1077612390',display_name='tempest-LiveMigrationTest-server-1077612390',ec2_ids=,ephemeral_gb=0,ephemeral_key_uuid=None,fault=,flavor=Flavor(5),hidden=False,host='np0005537446.localdomain',hostname='tempest-livemigrationtest-server-1077612390',id=8,image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',info_cache=InstanceInfoCache,instance_type_id=5,kernel_id='',key_data=None,key_name=None,keypairs=,launch_index=0,launched_at=2025-11-27T10:04:13Z,launched_on='np0005537446.localdomain',locked=False,locked_by=None,memory_mb=128,metadata={},migration_context=None,new_flavor=None,node='np0005537446.localdomain',numa_topology=None,old_flavor=None,os_type=None,pci_devices=,pci_requests=InstancePCIRequests,power_state=1,progress=0,project_id='a6d84e90a5fa4f74afa4103832d90ac2',ramdisk_id='',reservation_id='r-mchq7ras',resources=None,root_device_name='/dev/vda',root_gb=1,security_groups=SecurityGroupList,services=,shutdown_terminate=False,system_metadata={boot_roles='reader,member',image_base_image_ref='a85008d0-4cdf-44ee-935d-c55441a5aa59',image_container_format='bare',image_disk_format='qcow2',image_hw_cdrom_bus='sata',image_hw_disk_bus='virtio',image_hw_input_bus='usb',image_hw_machine_type='q35',image_hw_pointer_model='usbtablet',image_hw_rng_model='virtio',image_hw_video_model='virtio',image_hw_vif_model='virtio',image_min_disk='1',image_min_ram='0',owner_project_name='tempest-LiveMigrationTest-1823599681',owner_user_name='tempest-LiveMigrationTest-1823599681-project-member'},tags=,task_state='migrating',terminated_at=None,trusted_certs=,updated_at=2025-11-27T10:04:15Z,user_data=None,user_id='edb03acaf09647f6b6b5465b16878a77',uuid=765bb3f2-7b81-4d43-8eea-4f8b0e498677,vcpu_model=,vcpus=1,vm_mode=None,vm_state='active') vif={"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} unplug /usr/lib/python3.9/site-packages/nova/virt/libvirt/vif.py:828#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.383 284030 DEBUG nova.network.os_vif_util [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Converting VIF {"id": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "address": "fa:16:3e:75:1f:c2", "network": {"id": "2247f41d-32cc-48a7-83ff-88f84e00e09c", "bridge": "br-int", "label": "tempest-LiveMigrationTest-2003408591-network", "subnets": [{"cidr": "10.100.0.0/28", "dns": [], "gateway": {"address": "10.100.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "10.100.0.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "10.100.0.3"}}], "meta": {"injected": false, "tenant_id": "a6d84e90a5fa4f74afa4103832d90ac2", "mtu": 1442, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "bridge_name": "br-int", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tap20252d49-04", "ovs_interfaceid": "20252d49-04fc-4922-8fc1-99b7e3b77a67", "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}} nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:511#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.384 284030 DEBUG nova.network.os_vif_util [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Converted object VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') nova_to_osvif_vif /usr/lib/python3.9/site-packages/nova/network/os_vif_util.py:548#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.385 284030 DEBUG os_vif [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Unplugging vif VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04') unplug /usr/lib/python3.9/site-packages/os_vif/__init__.py:109#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.389 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 18 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.389 284030 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DelPortCommand(_result=None, port=tap20252d49-04, bridge=br-int, if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.397 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.400 284030 INFO os_vif [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Successfully unplugged vif VIFOpenVSwitch(active=False,address=fa:16:3e:75:1f:c2,bridge_name='br-int',has_traffic_filtering=True,id=20252d49-04fc-4922-8fc1-99b7e3b77a67,network=Network(2247f41d-32cc-48a7-83ff-88f84e00e09c),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=True,vif_name='tap20252d49-04')#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.401 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.free_pci_device_allocations_for_instance" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.401 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.free_pci_device_allocations_for_instance" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.402 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.free_pci_device_allocations_for_instance" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.402 284030 DEBUG nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Calling driver.cleanup from _post_live_migration _post_live_migration /usr/lib/python3.9/site-packages/nova/compute/manager.py:9349#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.403 284030 INFO nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Deleting instance files /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677_del#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.405 284030 INFO nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Deletion of /var/lib/nova/instances/765bb3f2-7b81-4d43-8eea-4f8b0e498677_del complete#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.453 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:25 localhost nova_compute[284026]: 2025-11-27 10:04:25.563 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:25 localhost openstack_network_exporter[244641]: ERROR 10:04:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:04:25 localhost openstack_network_exporter[244641]: ERROR 10:04:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:04:25 localhost openstack_network_exporter[244641]: ERROR 10:04:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:04:25 localhost openstack_network_exporter[244641]: ERROR 10:04:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:04:25 localhost openstack_network_exporter[244641]: Nov 27 05:04:25 localhost openstack_network_exporter[244641]: ERROR 10:04:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:04:25 localhost openstack_network_exporter[244641]: Nov 27 05:04:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:04:25 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:04:26 localhost podman[313427]: 2025-11-27 10:04:26.015643542 +0000 UTC m=+0.103739051 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 05:04:26 localhost podman[313427]: 2025-11-27 10:04:26.050939698 +0000 UTC m=+0.139035157 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:04:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e104 e104: 6 total, 6 up, 6 in Nov 27 05:04:26 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:04:26 localhost podman[313428]: 2025-11-27 10:04:26.114866199 +0000 UTC m=+0.201215081 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}) Nov 27 05:04:26 localhost podman[313428]: 2025-11-27 10:04:26.154727139 +0000 UTC m=+0.241076021 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd) Nov 27 05:04:26 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:04:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e105 e105: 6 total, 6 up, 6 in Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.761 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.763 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:04:26 localhost nova_compute[284026]: 2025-11-27 10:04:26.763 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:27.082 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '11'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:04:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v161: 177 pgs: 177 active+clean; 271 MiB data, 934 MiB used, 41 GiB / 42 GiB avail; 91 KiB/s rd, 9.5 KiB/s wr, 122 op/s Nov 27 05:04:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:04:27 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1123286623' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.233 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.469s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.308 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.308 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:04:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e105 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.581 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.583 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11246MB free_disk=41.6374626159668GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.583 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.583 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e106 e106: 6 total, 6 up, 6 in Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.660 284030 INFO nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Updating resource usage from migration 29b259ea-ef69-4cbd-be42-1ec4a8fedcae#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.706 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.706 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Migration 29b259ea-ef69-4cbd-be42-1ec4a8fedcae is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 128, 'DISK_GB': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1640#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.707 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 2 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.707 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1152MB phys_disk=41GB used_disk=3GB total_vcpus=8 used_vcpus=2 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:04:27 localhost nova_compute[284026]: 2025-11-27 10:04:27.796 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:04:28 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2163527114' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.304 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.508s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.312 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.329 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.356 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.357 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.773s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.932 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Acquiring lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.933 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.933 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "765bb3f2-7b81-4d43-8eea-4f8b0e498677-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.950 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.950 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.951 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.951 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:04:28 localhost nova_compute[284026]: 2025-11-27 10:04:28.951 284030 DEBUG oslo_concurrency.processutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v163: 177 pgs: 177 active+clean; 271 MiB data, 934 MiB used, 41 GiB / 42 GiB avail; 92 KiB/s rd, 9.6 KiB/s wr, 123 op/s Nov 27 05:04:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e107 e107: 6 total, 6 up, 6 in Nov 27 05:04:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:04:29 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3300680795' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.410 284030 DEBUG oslo_concurrency.processutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.459s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.464 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.464 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.650 284030 WARNING nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.652 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11257MB free_disk=41.6374626159668GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.652 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.653 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.695 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Migration for instance 765bb3f2-7b81-4d43-8eea-4f8b0e498677 refers to another host's instance! _pair_instances_to_migrations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:903#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.716 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Skipping migration as instance is neither resizing nor live-migrating. _update_usage_from_migrations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1491#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.763 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.764 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Migration 29b259ea-ef69-4cbd-be42-1ec4a8fedcae is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 128, 'DISK_GB': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1640#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.765 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:04:29 localhost nova_compute[284026]: 2025-11-27 10:04:29.765 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.086 284030 DEBUG oslo_concurrency.processutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:04:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e108 e108: 6 total, 6 up, 6 in Nov 27 05:04:30 localhost systemd[1]: tmp-crun.TRxcZm.mount: Deactivated successfully. Nov 27 05:04:30 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:04:30 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:30 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:30 localhost podman[313547]: 2025-11-27 10:04:30.285622123 +0000 UTC m=+0.084244693 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.357 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.358 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.358 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.565 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:04:30 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/413790241' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.615 284030 DEBUG oslo_concurrency.processutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.529s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.621 284030 DEBUG nova.compute.provider_tree [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:04:30 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:30.694 2 INFO neutron.agent.securitygroups_rpc [None req-a3a3956c-e390-47ce-8c07-7551a0725ccf 61f51f1c53ad41ed87ccfe3659c1cd65 21f82b6d03cf4cdc83a0702e249302d9 - - default default] Security group member updated ['7c7604d9-10e6-43d7-bd2c-14b42cf8e824']#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.699 284030 DEBUG nova.scheduler.client.report [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:04:30 localhost ovn_controller[156436]: 2025-11-27T10:04:30Z|00124|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.837 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.837 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.838 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.838 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.869 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.925 284030 DEBUG nova.compute.resource_tracker [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.926 284030 DEBUG oslo_concurrency.lockutils [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.273s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:30 localhost nova_compute[284026]: 2025-11-27 10:04:30.934 284030 INFO nova.compute.manager [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Migrating instance to np0005537444.localdomain finished successfully.#033[00m Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.043 284030 INFO nova.scheduler.client.report [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] Deleted allocation for migration 29b259ea-ef69-4cbd-be42-1ec4a8fedcae#033[00m Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.043 284030 DEBUG nova.virt.libvirt.driver [None req-bafc4dbe-ce13-4465-830a-532d7044dfde 8d2a50bb265d48deb5d191407d799b98 2f8a8262d8074ffd876910fbc25620fe - - default default] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Live migration monitoring is all done _live_migration /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:10662#033[00m Nov 27 05:04:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v166: 177 pgs: 177 active+clean; 306 MiB data, 1017 MiB used, 41 GiB / 42 GiB avail; 2.4 MiB/s rd, 5.8 MiB/s wr, 516 op/s Nov 27 05:04:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e109 e109: 6 total, 6 up, 6 in Nov 27 05:04:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e110 e110: 6 total, 6 up, 6 in Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.751 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.776 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.776 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:04:31 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:04:31 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:31 localhost podman[313606]: 2025-11-27 10:04:31.905730397 +0000 UTC m=+0.056673106 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 05:04:31 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:31 localhost systemd[1]: tmp-crun.EufBOe.mount: Deactivated successfully. Nov 27 05:04:31 localhost ovn_controller[156436]: 2025-11-27T10:04:31Z|00125|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:31.967 2 INFO neutron.agent.securitygroups_rpc [None req-23861d8b-cea5-4a79-b66b-6b509c0d70aa 61f51f1c53ad41ed87ccfe3659c1cd65 21f82b6d03cf4cdc83a0702e249302d9 - - default default] Security group member updated ['7c7604d9-10e6-43d7-bd2c-14b42cf8e824']#033[00m Nov 27 05:04:31 localhost nova_compute[284026]: 2025-11-27 10:04:31.979 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e110 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e111 e111: 6 total, 6 up, 6 in Nov 27 05:04:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:32.808 2 INFO neutron.agent.securitygroups_rpc [None req-822465e4-7546-44a0-ab0e-6afce6c1a3e8 61f51f1c53ad41ed87ccfe3659c1cd65 21f82b6d03cf4cdc83a0702e249302d9 - - default default] Security group member updated ['7c7604d9-10e6-43d7-bd2c-14b42cf8e824']#033[00m Nov 27 05:04:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v170: 177 pgs: 177 active+clean; 306 MiB data, 1017 MiB used, 41 GiB / 42 GiB avail; 2.7 MiB/s rd, 6.5 MiB/s wr, 582 op/s Nov 27 05:04:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:34.535 2 INFO neutron.agent.securitygroups_rpc [None req-62e064c7-f42a-4fc4-b902-29b793311568 61f51f1c53ad41ed87ccfe3659c1cd65 21f82b6d03cf4cdc83a0702e249302d9 - - default default] Security group member updated ['7c7604d9-10e6-43d7-bd2c-14b42cf8e824']#033[00m Nov 27 05:04:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e112 e112: 6 total, 6 up, 6 in Nov 27 05:04:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:04:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:04:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:04:35 localhost podman[313628]: 2025-11-27 10:04:34.987602767 +0000 UTC m=+0.081353895 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:04:35 localhost podman[313635]: 2025-11-27 10:04:35.056666787 +0000 UTC m=+0.137629249 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.buildah.version=1.33.7, managed_by=edpm_ansible, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, build-date=2025-08-20T13:12:41, distribution-scope=public, release=1755695350, com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:04:35 localhost podman[313635]: 2025-11-27 10:04:35.067823759 +0000 UTC m=+0.148786271 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, architecture=x86_64, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, managed_by=edpm_ansible, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, io.openshift.expose-services=, vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., name=ubi9-minimal, build-date=2025-08-20T13:12:41, distribution-scope=public, vcs-type=git) Nov 27 05:04:35 localhost podman[313628]: 2025-11-27 10:04:35.079001752 +0000 UTC m=+0.172752900 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, tcib_managed=true, config_id=edpm, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:35 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:04:35 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:04:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v172: 177 pgs: 177 active+clean; 306 MiB data, 1017 MiB used, 41 GiB / 42 GiB avail Nov 27 05:04:35 localhost podman[313629]: 2025-11-27 10:04:35.019238864 +0000 UTC m=+0.101103610 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:04:35 localhost podman[313629]: 2025-11-27 10:04:35.157118248 +0000 UTC m=+0.238983004 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:04:35 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:04:35 localhost nova_compute[284026]: 2025-11-27 10:04:35.428 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:35 localhost nova_compute[284026]: 2025-11-27 10:04:35.569 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e113 e113: 6 total, 6 up, 6 in Nov 27 05:04:35 localhost systemd[1]: tmp-crun.nP3APm.mount: Deactivated successfully. Nov 27 05:04:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v174: 177 pgs: 177 active+clean; 307 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 8.6 MiB/s rd, 8.5 MiB/s wr, 270 op/s Nov 27 05:04:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e113 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:38 localhost nova_compute[284026]: 2025-11-27 10:04:38.235 284030 DEBUG nova.virt.driver [-] Emitting event Stopped> emit_event /usr/lib/python3.9/site-packages/nova/virt/driver.py:1653#033[00m Nov 27 05:04:38 localhost nova_compute[284026]: 2025-11-27 10:04:38.235 284030 INFO nova.compute.manager [-] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] VM Stopped (Lifecycle Event)#033[00m Nov 27 05:04:38 localhost nova_compute[284026]: 2025-11-27 10:04:38.445 284030 DEBUG nova.compute.manager [None req-5202e78b-4f03-4e78-a51a-6b5d2716c147 - - - - - -] [instance: 765bb3f2-7b81-4d43-8eea-4f8b0e498677] Checking state _get_power_state /usr/lib/python3.9/site-packages/nova/compute/manager.py:1762#033[00m Nov 27 05:04:38 localhost podman[242678]: time="2025-11-27T10:04:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:04:38 localhost podman[242678]: @ - - [27/Nov/2025:10:04:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 161158 "" "Go-http-client/1.1" Nov 27 05:04:38 localhost podman[242678]: @ - - [27/Nov/2025:10:04:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20692 "" "Go-http-client/1.1" Nov 27 05:04:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v175: 177 pgs: 177 active+clean; 307 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 7.3 MiB/s rd, 7.2 MiB/s wr, 230 op/s Nov 27 05:04:39 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:39.741 2 INFO neutron.agent.securitygroups_rpc [None req-42916fd0-560f-4c2d-886a-1bd0e8438ef4 edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Security group member updated ['6d11bb1f-93da-41d7-8007-c66992d17e28']#033[00m Nov 27 05:04:40 localhost nova_compute[284026]: 2025-11-27 10:04:40.431 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:40 localhost dnsmasq[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/addn_hosts - 0 addresses Nov 27 05:04:40 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/host Nov 27 05:04:40 localhost podman[313708]: 2025-11-27 10:04:40.560969154 +0000 UTC m=+0.060175230 container kill 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:40 localhost dnsmasq-dhcp[312330]: read /var/lib/neutron/dhcp/14034b34-9875-46a0-b548-1a34b924b9b9/opts Nov 27 05:04:40 localhost nova_compute[284026]: 2025-11-27 10:04:40.572 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v176: 177 pgs: 177 active+clean; 226 MiB data, 906 MiB used, 41 GiB / 42 GiB avail; 6.0 MiB/s rd, 5.9 MiB/s wr, 248 op/s Nov 27 05:04:41 localhost dnsmasq[312330]: exiting on receipt of SIGTERM Nov 27 05:04:41 localhost podman[313745]: 2025-11-27 10:04:41.444438206 +0000 UTC m=+0.059933115 container kill 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:04:41 localhost systemd[1]: libpod-40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0.scope: Deactivated successfully. Nov 27 05:04:41 localhost podman[313759]: 2025-11-27 10:04:41.523043525 +0000 UTC m=+0.057663653 container died 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:04:41 localhost podman[313759]: 2025-11-27 10:04:41.556721837 +0000 UTC m=+0.091341935 container cleanup 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:04:41 localhost systemd[1]: var-lib-containers-storage-overlay-f4836a40e4a7d6eec8090c62b45f3fa53f77dab15c629bec5b2838dad003dfdb-merged.mount: Deactivated successfully. Nov 27 05:04:41 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:41 localhost systemd[1]: libpod-conmon-40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0.scope: Deactivated successfully. Nov 27 05:04:41 localhost podman[313760]: 2025-11-27 10:04:41.596771102 +0000 UTC m=+0.130097166 container remove 40835984ee6ca70aea507c6a1c70784eb67243983cef61d19bce93914e360fe0 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-14034b34-9875-46a0-b548-1a34b924b9b9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:04:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e114 e114: 6 total, 6 up, 6 in Nov 27 05:04:41 localhost ovn_controller[156436]: 2025-11-27T10:04:41Z|00126|binding|INFO|Releasing lport 7d36709c-2fe6-487c-9486-157a966792a0 from this chassis (sb_readonly=0) Nov 27 05:04:41 localhost kernel: device tap7d36709c-2f left promiscuous mode Nov 27 05:04:41 localhost nova_compute[284026]: 2025-11-27 10:04:41.644 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:41 localhost ovn_controller[156436]: 2025-11-27T10:04:41Z|00127|binding|INFO|Setting lport 7d36709c-2fe6-487c-9486-157a966792a0 down in Southbound Nov 27 05:04:41 localhost nova_compute[284026]: 2025-11-27 10:04:41.666 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:41 localhost nova_compute[284026]: 2025-11-27 10:04:41.667 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:42.037 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '19.80.0.3/24', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-14034b34-9875-46a0-b548-1a34b924b9b9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'a6d84e90a5fa4f74afa4103832d90ac2', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=d463cc2b-c33c-4d8e-8d92-b80cf5ac63e7, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=7d36709c-2fe6-487c-9486-157a966792a0) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:42.039 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 7d36709c-2fe6-487c-9486-157a966792a0 in datapath 14034b34-9875-46a0-b548-1a34b924b9b9 unbound from our chassis#033[00m Nov 27 05:04:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:42.041 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 14034b34-9875-46a0-b548-1a34b924b9b9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:42.043 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[509c7a6d-6e80-45c5-b3de-cb9550589c84]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:42 localhost systemd[1]: run-netns-qdhcp\x2d14034b34\x2d9875\x2d46a0\x2db548\x2d1a34b924b9b9.mount: Deactivated successfully. Nov 27 05:04:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e114 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:42.457 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:04:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v178: 177 pgs: 177 active+clean; 226 MiB data, 906 MiB used, 41 GiB / 42 GiB avail; 6.0 MiB/s rd, 5.9 MiB/s wr, 248 op/s Nov 27 05:04:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:43.296 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:04:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:43.573 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:04:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:43.574 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:04:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:43.575 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:04:43 localhost ovn_controller[156436]: 2025-11-27T10:04:43Z|00128|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:43.703 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:04:43Z, description=, device_id=6d5b6f14-dd29-4750-92b5-e0940f2e5fc1, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=914fd334-4649-44eb-a230-aed7aa95bb07, ip_allocation=immediate, mac_address=fa:16:3e:9b:dd:f8, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=810, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:04:43Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:04:43 localhost nova_compute[284026]: 2025-11-27 10:04:43.731 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:04:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:04:43 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:04:43 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:43 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:43 localhost podman[313808]: 2025-11-27 10:04:43.966903353 +0000 UTC m=+0.098176460 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:04:44 localhost podman[313819]: 2025-11-27 10:04:44.028700317 +0000 UTC m=+0.120009092 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:04:44 localhost podman[313819]: 2025-11-27 10:04:44.073905611 +0000 UTC m=+0.165214386 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:04:44 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:04:44 localhost podman[313818]: 2025-11-27 10:04:44.097606554 +0000 UTC m=+0.193387180 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, container_name=ovn_controller, config_id=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:04:44 localhost podman[313818]: 2025-11-27 10:04:44.131847761 +0000 UTC m=+0.227628407 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:04:44 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:04:44 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:44.272 265123 INFO neutron.agent.dhcp.agent [None req-784d5c93-7f0b-46de-859b-ed40f879e212 - - - - - -] DHCP configuration for ports {'914fd334-4649-44eb-a230-aed7aa95bb07'} is completed#033[00m Nov 27 05:04:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:04:44 Nov 27 05:04:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:04:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:04:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['manila_metadata', 'volumes', '.mgr', 'backups', 'images', 'vms', 'manila_data'] Nov 27 05:04:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:04:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:04:44.465 2 INFO neutron.agent.securitygroups_rpc [None req-bf4f8853-e733-4e71-9d73-0e60dab09faa edb03acaf09647f6b6b5465b16878a77 a6d84e90a5fa4f74afa4103832d90ac2 - - default default] Security group member updated ['6d11bb1f-93da-41d7-8007-c66992d17e28']#033[00m Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:04:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost snmpd[71019]: empty variable list in _query Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.0075513199678950305 of space, bias 1.0, pg target 1.507746886923041 quantized to 32 (current 32) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:04:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.0021628687418574354 quantized to 16 (current 16) Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:04:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:04:44 localhost nova_compute[284026]: 2025-11-27 10:04:44.812 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v179: 177 pgs: 177 active+clean; 226 MiB data, 906 MiB used, 41 GiB / 42 GiB avail; 38 KiB/s rd, 2.0 KiB/s wr, 53 op/s Nov 27 05:04:45 localhost nova_compute[284026]: 2025-11-27 10:04:45.434 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:45 localhost nova_compute[284026]: 2025-11-27 10:04:45.575 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v180: 177 pgs: 177 active+clean; 307 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 4.7 MiB/s rd, 4.7 MiB/s wr, 141 op/s Nov 27 05:04:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e114 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e115 e115: 6 total, 6 up, 6 in Nov 27 05:04:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:04:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/790857051' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:04:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:04:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/790857051' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:04:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v182: 177 pgs: 177 active+clean; 307 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 5.9 MiB/s rd, 5.8 MiB/s wr, 114 op/s Nov 27 05:04:49 localhost nova_compute[284026]: 2025-11-27 10:04:49.657 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:49 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:04:49 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:49 localhost podman[313888]: 2025-11-27 10:04:49.722095385 +0000 UTC m=+0.085352522 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 05:04:49 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:49 localhost systemd[1]: tmp-crun.4AebT0.mount: Deactivated successfully. Nov 27 05:04:49 localhost ovn_controller[156436]: 2025-11-27T10:04:49Z|00129|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:50 localhost nova_compute[284026]: 2025-11-27 10:04:50.024 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:50 localhost nova_compute[284026]: 2025-11-27 10:04:50.436 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:50 localhost nova_compute[284026]: 2025-11-27 10:04:50.577 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v183: 177 pgs: 177 active+clean; 226 MiB data, 887 MiB used, 41 GiB / 42 GiB avail; 7.4 MiB/s rd, 4.9 MiB/s wr, 219 op/s Nov 27 05:04:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e116 e116: 6 total, 6 up, 6 in Nov 27 05:04:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e116 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v185: 177 pgs: 177 active+clean; 226 MiB data, 887 MiB used, 41 GiB / 42 GiB avail; 8.8 MiB/s rd, 5.8 MiB/s wr, 259 op/s Nov 27 05:04:53 localhost dnsmasq[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/addn_hosts - 0 addresses Nov 27 05:04:53 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/host Nov 27 05:04:53 localhost dnsmasq-dhcp[311647]: read /var/lib/neutron/dhcp/75d79a30-b416-43bb-83b3-cd22cce0e128/opts Nov 27 05:04:53 localhost podman[313925]: 2025-11-27 10:04:53.621876451 +0000 UTC m=+0.072811263 container kill 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:04:53 localhost systemd[1]: tmp-crun.C7u85d.mount: Deactivated successfully. Nov 27 05:04:54 localhost nova_compute[284026]: 2025-11-27 10:04:54.023 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:54 localhost ovn_controller[156436]: 2025-11-27T10:04:54Z|00130|binding|INFO|Releasing lport ae33bdba-c03c-4d06-a040-bfb31931b448 from this chassis (sb_readonly=0) Nov 27 05:04:54 localhost kernel: device tapae33bdba-c0 left promiscuous mode Nov 27 05:04:54 localhost ovn_controller[156436]: 2025-11-27T10:04:54Z|00131|binding|INFO|Setting lport ae33bdba-c03c-4d06-a040-bfb31931b448 down in Southbound Nov 27 05:04:54 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:54.034 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.3/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-75d79a30-b416-43bb-83b3-cd22cce0e128', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-75d79a30-b416-43bb-83b3-cd22cce0e128', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '2f8a8262d8074ffd876910fbc25620fe', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=6fde6c61-2676-4137-9e99-778e99b077f8, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=ae33bdba-c03c-4d06-a040-bfb31931b448) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:54 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:54.036 162092 INFO neutron.agent.ovn.metadata.agent [-] Port ae33bdba-c03c-4d06-a040-bfb31931b448 in datapath 75d79a30-b416-43bb-83b3-cd22cce0e128 unbound from our chassis#033[00m Nov 27 05:04:54 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:54.039 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 75d79a30-b416-43bb-83b3-cd22cce0e128, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:04:54 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:54.041 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3b3a75ef-7526-49d9-9cdf-704515cb95e3]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:04:54 localhost nova_compute[284026]: 2025-11-27 10:04:54.046 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v186: 177 pgs: 177 active+clean; 226 MiB data, 887 MiB used, 41 GiB / 42 GiB avail; 2.9 MiB/s rd, 20 KiB/s wr, 145 op/s Nov 27 05:04:55 localhost nova_compute[284026]: 2025-11-27 10:04:55.439 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:55 localhost nova_compute[284026]: 2025-11-27 10:04:55.617 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:55 localhost openstack_network_exporter[244641]: ERROR 10:04:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:04:55 localhost openstack_network_exporter[244641]: ERROR 10:04:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:04:55 localhost openstack_network_exporter[244641]: ERROR 10:04:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:04:55 localhost openstack_network_exporter[244641]: ERROR 10:04:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:04:55 localhost openstack_network_exporter[244641]: Nov 27 05:04:55 localhost openstack_network_exporter[244641]: ERROR 10:04:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:04:55 localhost openstack_network_exporter[244641]: Nov 27 05:04:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e117 e117: 6 total, 6 up, 6 in Nov 27 05:04:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:04:56 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:04:57 localhost systemd[1]: tmp-crun.itwnmF.mount: Deactivated successfully. Nov 27 05:04:57 localhost podman[313948]: 2025-11-27 10:04:57.007351545 +0000 UTC m=+0.103417402 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:04:57 localhost podman[313948]: 2025-11-27 10:04:57.018059036 +0000 UTC m=+0.114124903 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_managed=true) Nov 27 05:04:57 localhost systemd[1]: tmp-crun.UiX6Np.mount: Deactivated successfully. Nov 27 05:04:57 localhost podman[313949]: 2025-11-27 10:04:57.062172341 +0000 UTC m=+0.147665121 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:57 localhost podman[313949]: 2025-11-27 10:04:57.074935876 +0000 UTC m=+0.160428696 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:04:57 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:04:57 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:04:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v188: 177 pgs: 177 active+clean; 145 MiB data, 746 MiB used, 41 GiB / 42 GiB avail; 2.9 MiB/s rd, 22 KiB/s wr, 186 op/s Nov 27 05:04:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e117 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:04:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:04:58 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:04:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:04:58 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:04:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:04:58 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev ab9a2a76-4963-4f19-8d34-62a130eeb883 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:04:58 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev ab9a2a76-4963-4f19-8d34-62a130eeb883 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:04:58 localhost ceph-mgr[290377]: [progress INFO root] Completed event ab9a2a76-4963-4f19-8d34-62a130eeb883 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:04:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:04:58 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:04:58 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:04:58 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:58 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:58 localhost podman[314085]: 2025-11-27 10:04:58.544241276 +0000 UTC m=+0.066953995 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:04:58 localhost ovn_controller[156436]: 2025-11-27T10:04:58Z|00132|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:04:58 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:04:58 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:04:58 localhost nova_compute[284026]: 2025-11-27 10:04:58.757 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e118 e118: 6 total, 6 up, 6 in Nov 27 05:04:58 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:04:58 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:04:58 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:04:58 localhost systemd[1]: tmp-crun.QN2HRJ.mount: Deactivated successfully. Nov 27 05:04:58 localhost podman[314127]: 2025-11-27 10:04:58.925645217 +0000 UTC m=+0.056983665 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:04:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v190: 177 pgs: 177 active+clean; 145 MiB data, 746 MiB used, 41 GiB / 42 GiB avail; 30 KiB/s rd, 1.9 KiB/s wr, 43 op/s Nov 27 05:04:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:59.206 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=12, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=11) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:04:59 localhost nova_compute[284026]: 2025-11-27 10:04:59.207 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:04:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:04:59.207 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 7 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:04:59 localhost dnsmasq[311647]: exiting on receipt of SIGTERM Nov 27 05:04:59 localhost podman[314165]: 2025-11-27 10:04:59.555796626 +0000 UTC m=+0.063311246 container kill 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:59 localhost systemd[1]: libpod-51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f.scope: Deactivated successfully. Nov 27 05:04:59 localhost podman[314180]: 2025-11-27 10:04:59.632259167 +0000 UTC m=+0.057174279 container died 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:04:59 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:04:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:04:59 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f-userdata-shm.mount: Deactivated successfully. Nov 27 05:04:59 localhost systemd[1]: var-lib-containers-storage-overlay-14f73f1e31358fea30c828fc2907fbab9fa54e1cddf6e2ea4115fbbf651e33c4-merged.mount: Deactivated successfully. Nov 27 05:04:59 localhost podman[314180]: 2025-11-27 10:04:59.686624929 +0000 UTC m=+0.111539991 container remove 51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-75d79a30-b416-43bb-83b3-cd22cce0e128, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:04:59 localhost systemd[1]: libpod-conmon-51add4c9ce3740f9db649df3338e820293ca4ba04f50d4875d3b1c53ed38dc9f.scope: Deactivated successfully. Nov 27 05:04:59 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:04:59 localhost systemd[1]: run-netns-qdhcp\x2d75d79a30\x2db416\x2d43bb\x2d83b3\x2dcd22cce0e128.mount: Deactivated successfully. Nov 27 05:04:59 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:04:59.892 265123 INFO neutron.agent.dhcp.agent [None req-d4bce002-d52d-4ca8-a9f2-126158091002 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:00.127 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:00 localhost nova_compute[284026]: 2025-11-27 10:05:00.443 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:00 localhost nova_compute[284026]: 2025-11-27 10:05:00.616 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e119 e119: 6 total, 6 up, 6 in Nov 27 05:05:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:01.084 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v192: 177 pgs: 177 active+clean; 145 MiB data, 764 MiB used, 41 GiB / 42 GiB avail; 132 KiB/s rd, 12 KiB/s wr, 185 op/s Nov 27 05:05:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e119 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e120 e120: 6 total, 6 up, 6 in Nov 27 05:05:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v194: 177 pgs: 177 active+clean; 145 MiB data, 764 MiB used, 41 GiB / 42 GiB avail; 94 KiB/s rd, 10 KiB/s wr, 130 op/s Nov 27 05:05:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v195: 177 pgs: 177 active+clean; 145 MiB data, 764 MiB used, 41 GiB / 42 GiB avail; 89 KiB/s rd, 9.6 KiB/s wr, 123 op/s Nov 27 05:05:05 localhost nova_compute[284026]: 2025-11-27 10:05:05.447 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:05 localhost nova_compute[284026]: 2025-11-27 10:05:05.658 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:05 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:05.661 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:05:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:05:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:05:06 localhost systemd[1]: tmp-crun.Kq2Br9.mount: Deactivated successfully. Nov 27 05:05:06 localhost podman[314205]: 2025-11-27 10:05:06.015794361 +0000 UTC m=+0.104826541 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:05:06 localhost podman[314206]: 2025-11-27 10:05:06.049949496 +0000 UTC m=+0.135945844 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:05:06 localhost podman[314206]: 2025-11-27 10:05:06.062060524 +0000 UTC m=+0.148056922 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:05:06 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:05:06 localhost podman[314205]: 2025-11-27 10:05:06.148697661 +0000 UTC m=+0.237729841 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, tcib_managed=true, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:05:06 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:05:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:06.209 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '12'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:05:06 localhost podman[314207]: 2025-11-27 10:05:06.214894744 +0000 UTC m=+0.295752753 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-type=git, version=9.6, architecture=x86_64, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, distribution-scope=public, maintainer=Red Hat, Inc., io.buildah.version=1.33.7, release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, container_name=openstack_network_exporter) Nov 27 05:05:06 localhost podman[314207]: 2025-11-27 10:05:06.227833674 +0000 UTC m=+0.308691673 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, managed_by=edpm_ansible, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, config_id=edpm, release=1755695350, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, name=ubi9-minimal, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, architecture=x86_64, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, vendor=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 05:05:06 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:05:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v196: 177 pgs: 177 active+clean; 145 MiB data, 768 MiB used, 41 GiB / 42 GiB avail; 123 KiB/s rd, 11 KiB/s wr, 169 op/s Nov 27 05:05:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e120 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:08.632 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:08Z, description=, device_id=9ecb218b-28e9-41fa-96b7-30eb10daa85f, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=031468ef-76b3-42eb-8c7e-ddde6ff7e461, ip_allocation=immediate, mac_address=fa:16:3e:d1:50:12, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=951, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:05:08Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:05:08 localhost podman[242678]: time="2025-11-27T10:05:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:05:08 localhost podman[242678]: @ - - [27/Nov/2025:10:05:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:05:08 localhost podman[242678]: @ - - [27/Nov/2025:10:05:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19729 "" "Go-http-client/1.1" Nov 27 05:05:08 localhost systemd[1]: tmp-crun.d6mwFo.mount: Deactivated successfully. Nov 27 05:05:08 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:05:08 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:08 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:08 localhost podman[314282]: 2025-11-27 10:05:08.92893103 +0000 UTC m=+0.064243511 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:05:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v197: 177 pgs: 177 active+clean; 145 MiB data, 768 MiB used, 41 GiB / 42 GiB avail; 50 KiB/s rd, 3.6 KiB/s wr, 68 op/s Nov 27 05:05:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:09.162 265123 INFO neutron.agent.dhcp.agent [None req-30246e88-db37-4c71-8e94-e2401a97451d - - - - - -] DHCP configuration for ports {'031468ef-76b3-42eb-8c7e-ddde6ff7e461'} is completed#033[00m Nov 27 05:05:10 localhost nova_compute[284026]: 2025-11-27 10:05:10.450 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:10 localhost nova_compute[284026]: 2025-11-27 10:05:10.661 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:10 localhost nova_compute[284026]: 2025-11-27 10:05:10.875 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v198: 177 pgs: 177 active+clean; 145 MiB data, 768 MiB used, 41 GiB / 42 GiB avail; 42 KiB/s rd, 3.0 KiB/s wr, 57 op/s Nov 27 05:05:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e121 e121: 6 total, 6 up, 6 in Nov 27 05:05:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e121 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v200: 177 pgs: 177 active+clean; 145 MiB data, 768 MiB used, 41 GiB / 42 GiB avail; 42 KiB/s rd, 3.0 KiB/s wr, 57 op/s Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:05:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:05:14 localhost podman[314302]: 2025-11-27 10:05:14.72904653 +0000 UTC m=+0.081957340 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, container_name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_controller) Nov 27 05:05:14 localhost podman[314303]: 2025-11-27 10:05:14.779418335 +0000 UTC m=+0.128505212 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:05:14 localhost podman[314302]: 2025-11-27 10:05:14.783925457 +0000 UTC m=+0.136836297 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:05:14 localhost podman[314303]: 2025-11-27 10:05:14.795781958 +0000 UTC m=+0.144868835 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:05:14 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:05:14 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:05:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v201: 177 pgs: 177 active+clean; 145 MiB data, 768 MiB used, 41 GiB / 42 GiB avail; 42 KiB/s rd, 3.0 KiB/s wr, 57 op/s Nov 27 05:05:15 localhost nova_compute[284026]: 2025-11-27 10:05:15.454 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:15 localhost nova_compute[284026]: 2025-11-27 10:05:15.664 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:16 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:16.117 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:15Z, description=, device_id=b2281776-df2e-4a94-825d-be3ee27f7234, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=4c6073cf-7b59-4dc4-b067-aa2752dc8c42, ip_allocation=immediate, mac_address=fa:16:3e:c5:e7:be, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1000, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:05:15Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:05:16 localhost nova_compute[284026]: 2025-11-27 10:05:16.187 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:16 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:05:16 localhost podman[314364]: 2025-11-27 10:05:16.342606648 +0000 UTC m=+0.066020500 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:05:16 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:16 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:16 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:16.564 265123 INFO neutron.agent.dhcp.agent [None req-8fbe2111-c3a8-49bc-8645-55f5f7eb788c - - - - - -] DHCP configuration for ports {'4c6073cf-7b59-4dc4-b067-aa2752dc8c42'} is completed#033[00m Nov 27 05:05:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v202: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e121 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:17 localhost nova_compute[284026]: 2025-11-27 10:05:17.576 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:18 localhost nova_compute[284026]: 2025-11-27 10:05:18.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v203: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:20 localhost nova_compute[284026]: 2025-11-27 10:05:20.457 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:20 localhost nova_compute[284026]: 2025-11-27 10:05:20.706 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:20 localhost nova_compute[284026]: 2025-11-27 10:05:20.730 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:20 localhost nova_compute[284026]: 2025-11-27 10:05:20.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v204: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:21 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:21.294 265123 INFO neutron.agent.linux.ip_lib [None req-d97f0cc9-6593-4da1-9911-3b32dd144ea8 - - - - - -] Device tapb53117af-13 cannot be used as it has no MAC address#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.316 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:21 localhost kernel: device tapb53117af-13 entered promiscuous mode Nov 27 05:05:21 localhost NetworkManager[5971]: [1764237921.3260] manager: (tapb53117af-13): new Generic device (/org/freedesktop/NetworkManager/Devices/27) Nov 27 05:05:21 localhost ovn_controller[156436]: 2025-11-27T10:05:21Z|00133|binding|INFO|Claiming lport b53117af-136a-4126-b12b-dbdf898758f5 for this chassis. Nov 27 05:05:21 localhost ovn_controller[156436]: 2025-11-27T10:05:21Z|00134|binding|INFO|b53117af-136a-4126-b12b-dbdf898758f5: Claiming unknown Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.327 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:21 localhost systemd-udevd[314396]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:21 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:21.336 2 INFO neutron.agent.securitygroups_rpc [None req-7f58d0b0-3267-4c75-bc8d-c21d31c298fb 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:21.339 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=b53117af-136a-4126-b12b-dbdf898758f5) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:21.341 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b53117af-136a-4126-b12b-dbdf898758f5 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:21.342 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:21.345 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0589f500-689e-4204-9518-579a7b01a5db]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:21 localhost ovn_controller[156436]: 2025-11-27T10:05:21Z|00135|binding|INFO|Setting lport b53117af-136a-4126-b12b-dbdf898758f5 ovn-installed in OVS Nov 27 05:05:21 localhost ovn_controller[156436]: 2025-11-27T10:05:21Z|00136|binding|INFO|Setting lport b53117af-136a-4126-b12b-dbdf898758f5 up in Southbound Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.375 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.407 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.441 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:21 localhost nova_compute[284026]: 2025-11-27 10:05:21.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:05:22 localhost podman[314451]: Nov 27 05:05:22 localhost podman[314451]: 2025-11-27 10:05:22.239942811 +0000 UTC m=+0.089021762 container create 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:05:22 localhost systemd[1]: Started libpod-conmon-743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a.scope. Nov 27 05:05:22 localhost systemd[1]: Started libcrun container. Nov 27 05:05:22 localhost podman[314451]: 2025-11-27 10:05:22.197379788 +0000 UTC m=+0.046458829 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:22 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/80c6e07108435863de99fc6e27664a8094d073d5ce35d8f6d4ce82768ea90123/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:22 localhost podman[314451]: 2025-11-27 10:05:22.307237814 +0000 UTC m=+0.156316815 container init 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:22 localhost podman[314451]: 2025-11-27 10:05:22.317985105 +0000 UTC m=+0.167064106 container start 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:05:22 localhost dnsmasq[314469]: started, version 2.85 cachesize 150 Nov 27 05:05:22 localhost dnsmasq[314469]: DNS service limited to local subnets Nov 27 05:05:22 localhost dnsmasq[314469]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:22 localhost dnsmasq[314469]: warning: no upstream servers configured Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:05:22 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e121 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:22.383 265123 INFO neutron.agent.dhcp.agent [None req-d97f0cc9-6593-4da1-9911-3b32dd144ea8 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:20Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=ab1a8000-2739-4bb5-aca6-13f36216b72e, ip_allocation=immediate, mac_address=fa:16:3e:c9:bf:b4, name=tempest-NetworksTestDHCPv6-1477293247, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['ae8f7087-e012-4345-9c09-22367e53aaa7'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:20Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1029, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:21Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:22.466 265123 INFO neutron.agent.dhcp.agent [None req-c8a9bca4-979c-49ca-887a-731e76727431 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:22 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:22.548 2 INFO neutron.agent.securitygroups_rpc [None req-7a492b7a-f7e1-4e4e-86f6-8dca5280ee54 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:22 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:22 localhost podman[314488]: 2025-11-27 10:05:22.580594129 +0000 UTC m=+0.059525244 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:05:22 localhost nova_compute[284026]: 2025-11-27 10:05:22.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:22.781 265123 INFO neutron.agent.dhcp.agent [None req-4a089525-ede2-44bd-a686-c53d655613e3 - - - - - -] DHCP configuration for ports {'ab1a8000-2739-4bb5-aca6-13f36216b72e'} is completed#033[00m Nov 27 05:05:22 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:22 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:22 localhost podman[314527]: 2025-11-27 10:05:22.897153444 +0000 UTC m=+0.054277322 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:05:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v205: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:23 localhost ovn_controller[156436]: 2025-11-27T10:05:23Z|00137|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:05:23 localhost nova_compute[284026]: 2025-11-27 10:05:23.514 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:23 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:05:23 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:23 localhost podman[314565]: 2025-11-27 10:05:23.604540944 +0000 UTC m=+0.054608279 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:05:23 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:23 localhost nova_compute[284026]: 2025-11-27 10:05:23.776 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:23 localhost ovn_controller[156436]: 2025-11-27T10:05:23Z|00138|binding|INFO|Releasing lport b7c183b8-b917-4950-913d-4efc5e8c5146 from this chassis (sb_readonly=0) Nov 27 05:05:23 localhost ovn_controller[156436]: 2025-11-27T10:05:23Z|00139|binding|INFO|Setting lport b7c183b8-b917-4950-913d-4efc5e8c5146 down in Southbound Nov 27 05:05:23 localhost kernel: device tapb7c183b8-b9 left promiscuous mode Nov 27 05:05:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:23.786 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '192.168.122.172/24', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-38b54160-bf0a-4463-b656-cfe2a3341789', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-38b54160-bf0a-4463-b656-cfe2a3341789', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '02a308de13ca4518beaddd402ba93936', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=4e3b70e4-6958-41d7-8c0c-f71821cf7e22, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=b7c183b8-b917-4950-913d-4efc5e8c5146) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:23.788 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b7c183b8-b917-4950-913d-4efc5e8c5146 in datapath 38b54160-bf0a-4463-b656-cfe2a3341789 unbound from our chassis#033[00m Nov 27 05:05:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:23.791 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 38b54160-bf0a-4463-b656-cfe2a3341789, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:05:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:23.792 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[9e4223c0-86a8-408c-88ff-806413849182]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:23 localhost nova_compute[284026]: 2025-11-27 10:05:23.800 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:23 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:23 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:23 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:23 localhost podman[314601]: 2025-11-27 10:05:23.805264192 +0000 UTC m=+0.066438931 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:24.160 265123 INFO neutron.agent.dhcp.agent [None req-0d4bd68a-d917-473a-8311-4f1cccd1b58d - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea', 'b53117af-136a-4126-b12b-dbdf898758f5'} is completed#033[00m Nov 27 05:05:24 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:24 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:24 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:24 localhost podman[314644]: 2025-11-27 10:05:24.254621723 +0000 UTC m=+0.043826287 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:05:24 localhost nova_compute[284026]: 2025-11-27 10:05:24.495 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:24 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:24.721 2 INFO neutron.agent.securitygroups_rpc [None req-ae685ab3-1717-4abc-8beb-e4351ad99893 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:24.825 265123 INFO neutron.agent.dhcp.agent [None req-522ca360-4974-48ed-bba2-e28c387fa93e - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea', 'b53117af-136a-4126-b12b-dbdf898758f5'} is completed#033[00m Nov 27 05:05:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:24.857 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:23Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=6cf7eac0-16fc-420b-a424-dbad9a67c0b3, ip_allocation=immediate, mac_address=fa:16:3e:5a:28:12, name=tempest-NetworksTestDHCPv6-826132186, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=4, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['795d658f-f1d7-455d-8bf6-7887f2cdfad6'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:23Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1042, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:24Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:25 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:25 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:25 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:25 localhost podman[314682]: 2025-11-27 10:05:25.024669892 +0000 UTC m=+0.050328184 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:05:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v206: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:25 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:25.273 265123 INFO neutron.agent.dhcp.agent [None req-0f814046-240a-4407-8c05-b1c18cbcad89 - - - - - -] DHCP configuration for ports {'6cf7eac0-16fc-420b-a424-dbad9a67c0b3'} is completed#033[00m Nov 27 05:05:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:25.425 2 INFO neutron.agent.securitygroups_rpc [None req-882f42c3-d7ce-4e24-8194-193e136bc270 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:25 localhost nova_compute[284026]: 2025-11-27 10:05:25.458 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:25 localhost dnsmasq[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:25 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:25 localhost dnsmasq-dhcp[314469]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:25 localhost podman[314721]: 2025-11-27 10:05:25.621626192 +0000 UTC m=+0.058117975 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:25 localhost nova_compute[284026]: 2025-11-27 10:05:25.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:25 localhost nova_compute[284026]: 2025-11-27 10:05:25.738 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:25 localhost nova_compute[284026]: 2025-11-27 10:05:25.749 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:25 localhost openstack_network_exporter[244641]: ERROR 10:05:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:05:25 localhost openstack_network_exporter[244641]: ERROR 10:05:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:05:25 localhost openstack_network_exporter[244641]: ERROR 10:05:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:05:25 localhost openstack_network_exporter[244641]: ERROR 10:05:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:05:25 localhost openstack_network_exporter[244641]: Nov 27 05:05:25 localhost openstack_network_exporter[244641]: ERROR 10:05:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:05:25 localhost openstack_network_exporter[244641]: Nov 27 05:05:26 localhost dnsmasq[314469]: exiting on receipt of SIGTERM Nov 27 05:05:26 localhost podman[314761]: 2025-11-27 10:05:26.27426962 +0000 UTC m=+0.060271523 container kill 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:05:26 localhost systemd[1]: libpod-743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a.scope: Deactivated successfully. Nov 27 05:05:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:26.310 2 INFO neutron.agent.securitygroups_rpc [req-bfe59756-5337-42f6-8a79-78c4b5a38bf9 req-38a2fea8-1ca1-4499-947d-45c92e2387f2 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['b7796825-0743-4837-a32d-48c32b9f68c2']#033[00m Nov 27 05:05:26 localhost podman[314773]: 2025-11-27 10:05:26.348919952 +0000 UTC m=+0.060467499 container died 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:05:26 localhost podman[314773]: 2025-11-27 10:05:26.389019238 +0000 UTC m=+0.100566745 container cleanup 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:26 localhost systemd[1]: libpod-conmon-743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a.scope: Deactivated successfully. Nov 27 05:05:26 localhost podman[314775]: 2025-11-27 10:05:26.426764911 +0000 UTC m=+0.131983186 container remove 743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:05:26 localhost nova_compute[284026]: 2025-11-27 10:05:26.439 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:26 localhost ovn_controller[156436]: 2025-11-27T10:05:26Z|00140|binding|INFO|Releasing lport b53117af-136a-4126-b12b-dbdf898758f5 from this chassis (sb_readonly=0) Nov 27 05:05:26 localhost ovn_controller[156436]: 2025-11-27T10:05:26Z|00141|binding|INFO|Setting lport b53117af-136a-4126-b12b-dbdf898758f5 down in Southbound Nov 27 05:05:26 localhost kernel: device tapb53117af-13 left promiscuous mode Nov 27 05:05:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:26.448 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=b53117af-136a-4126-b12b-dbdf898758f5) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:26.450 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b53117af-136a-4126-b12b-dbdf898758f5 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:26.452 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:26.453 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b6c85963-8849-43d7-adf4-c7fa99241682]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:26 localhost nova_compute[284026]: 2025-11-27 10:05:26.460 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:26 localhost systemd[1]: var-lib-containers-storage-overlay-80c6e07108435863de99fc6e27664a8094d073d5ce35d8f6d4ce82768ea90123-merged.mount: Deactivated successfully. Nov 27 05:05:26 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-743e6f3c7ee7fa11cb6183fee62cd6f29258c9b00a8d83e3969d42279c5d151a-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:26 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:26.748 265123 INFO neutron.agent.dhcp.agent [None req-e7bead53-63fc-4824-abf7-20e2cd95e94b - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v207: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e121 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:27.608 2 INFO neutron.agent.securitygroups_rpc [None req-b8069074-a997-43a4-aba5-3c2f3e5a2c2c 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:27.714 2 INFO neutron.agent.securitygroups_rpc [req-f48e6aab-c4ff-4854-8a5c-bfb653afb18b req-c131fedd-5c11-4db5-9f80-295466e0e4bf 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['2190fed3-2a24-4fa7-a2a4-fbae7a2031ce']#033[00m Nov 27 05:05:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:05:27 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:05:27 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:27.883 265123 INFO neutron.agent.linux.ip_lib [None req-529b5cff-2255-4aaf-a300-85b96835a968 - - - - - -] Device tap05855691-95 cannot be used as it has no MAC address#033[00m Nov 27 05:05:27 localhost systemd[1]: tmp-crun.L4wtYk.mount: Deactivated successfully. Nov 27 05:05:27 localhost podman[314805]: 2025-11-27 10:05:27.902708901 +0000 UTC m=+0.083291258 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 05:05:27 localhost nova_compute[284026]: 2025-11-27 10:05:27.911 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:27 localhost kernel: device tap05855691-95 entered promiscuous mode Nov 27 05:05:27 localhost NetworkManager[5971]: [1764237927.9198] manager: (tap05855691-95): new Generic device (/org/freedesktop/NetworkManager/Devices/28) Nov 27 05:05:27 localhost ovn_controller[156436]: 2025-11-27T10:05:27Z|00142|binding|INFO|Claiming lport 05855691-95b3-4737-b777-8cb539ffdd30 for this chassis. Nov 27 05:05:27 localhost ovn_controller[156436]: 2025-11-27T10:05:27Z|00143|binding|INFO|05855691-95b3-4737-b777-8cb539ffdd30: Claiming unknown Nov 27 05:05:27 localhost nova_compute[284026]: 2025-11-27 10:05:27.919 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:27 localhost systemd-udevd[314839]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:27.927 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=05855691-95b3-4737-b777-8cb539ffdd30) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:27.928 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 05855691-95b3-4737-b777-8cb539ffdd30 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:27.929 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:27.931 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0fb367df-d528-4052-b025-052ca63130fd]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:27 localhost podman[314805]: 2025-11-27 10:05:27.9418266 +0000 UTC m=+0.122408947 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:05:27 localhost ovn_controller[156436]: 2025-11-27T10:05:27Z|00144|binding|INFO|Setting lport 05855691-95b3-4737-b777-8cb539ffdd30 ovn-installed in OVS Nov 27 05:05:27 localhost ovn_controller[156436]: 2025-11-27T10:05:27Z|00145|binding|INFO|Setting lport 05855691-95b3-4737-b777-8cb539ffdd30 up in Southbound Nov 27 05:05:27 localhost nova_compute[284026]: 2025-11-27 10:05:27.944 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:27 localhost nova_compute[284026]: 2025-11-27 10:05:27.955 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost systemd[1]: tmp-crun.eYqqW9.mount: Deactivated successfully. Nov 27 05:05:27 localhost podman[314804]: 2025-11-27 10:05:27.960005682 +0000 UTC m=+0.143280342 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost podman[314804]: 2025-11-27 10:05:27.967197398 +0000 UTC m=+0.150472018 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:05:27 localhost journal[232028]: ethtool ioctl error on tap05855691-95: No such device Nov 27 05:05:27 localhost nova_compute[284026]: 2025-11-27 10:05:27.995 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.018 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.763 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.763 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.763 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:05:28 localhost nova_compute[284026]: 2025-11-27 10:05:28.764 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:05:28 localhost podman[314915]: Nov 27 05:05:28 localhost podman[314915]: 2025-11-27 10:05:28.787873878 +0000 UTC m=+0.095420917 container create 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:05:28 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:28.809 2 INFO neutron.agent.securitygroups_rpc [None req-b8debbfa-303c-40a6-9a72-57a2703d5197 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:28 localhost systemd[1]: Started libpod-conmon-62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e.scope. Nov 27 05:05:28 localhost podman[314915]: 2025-11-27 10:05:28.742808426 +0000 UTC m=+0.050355505 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:28 localhost systemd[1]: Started libcrun container. Nov 27 05:05:28 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/83e01c47ecdd532f5a55d83463f2c6b070c25b1662486e56988633d11c0eaff0/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:28 localhost podman[314915]: 2025-11-27 10:05:28.865705175 +0000 UTC m=+0.173252224 container init 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:05:28 localhost podman[314915]: 2025-11-27 10:05:28.882345336 +0000 UTC m=+0.189892375 container start 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:05:28 localhost dnsmasq[314936]: started, version 2.85 cachesize 150 Nov 27 05:05:28 localhost dnsmasq[314936]: DNS service limited to local subnets Nov 27 05:05:28 localhost dnsmasq[314936]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:28 localhost dnsmasq[314936]: warning: no upstream servers configured Nov 27 05:05:28 localhost dnsmasq-dhcp[314936]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:05:28 localhost dnsmasq[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:28 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:28 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:28.928 265123 INFO neutron.agent.dhcp.agent [None req-529b5cff-2255-4aaf-a300-85b96835a968 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:27Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=98fa4beb-beaf-48d8-b222-8798054e7a21, ip_allocation=immediate, mac_address=fa:16:3e:38:c8:77, name=tempest-NetworksTestDHCPv6-1882937899, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=6, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['46ef6b04-5e15-489a-b5ad-653c0cc92b20'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:26Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1067, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:27Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:29 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:29.098 265123 INFO neutron.agent.dhcp.agent [None req-77306d86-7252-4166-9209-2155af93780f - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:29 localhost dnsmasq[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:29 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:29 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:29 localhost podman[314973]: 2025-11-27 10:05:29.105630025 +0000 UTC m=+0.061018244 container kill 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:05:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v208: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:05:29 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2059084078' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.232 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.469s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.298 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.298 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:05:29 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:29.318 265123 INFO neutron.agent.dhcp.agent [None req-27ca6038-6433-4c09-b491-022c62afba4d - - - - - -] DHCP configuration for ports {'98fa4beb-beaf-48d8-b222-8798054e7a21'} is completed#033[00m Nov 27 05:05:29 localhost dnsmasq[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:29 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:29 localhost podman[315013]: 2025-11-27 10:05:29.420863783 +0000 UTC m=+0.062176525 container kill 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:05:29 localhost dnsmasq-dhcp[314936]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.463 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.464 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11250MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.464 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.465 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.569 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.570 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.570 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:05:29 localhost nova_compute[284026]: 2025-11-27 10:05:29.647 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:05:29 localhost systemd[1]: tmp-crun.GBJ1LS.mount: Deactivated successfully. Nov 27 05:05:29 localhost dnsmasq[314936]: exiting on receipt of SIGTERM Nov 27 05:05:29 localhost podman[315069]: 2025-11-27 10:05:29.878438568 +0000 UTC m=+0.057802137 container kill 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:29 localhost systemd[1]: libpod-62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e.scope: Deactivated successfully. Nov 27 05:05:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:29.900 2 INFO neutron.agent.securitygroups_rpc [req-e8f3dbcc-122d-4d58-9c3c-c2ffa4b89deb req-ea07783a-6644-4c3f-b8de-8ea9d5df3fbb 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['999d522a-887f-4a36-8ac4-991b3057ef22']#033[00m Nov 27 05:05:29 localhost podman[315082]: 2025-11-27 10:05:29.947381885 +0000 UTC m=+0.056336537 container died 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:29 localhost podman[315082]: 2025-11-27 10:05:29.981104388 +0000 UTC m=+0.090059000 container cleanup 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 05:05:29 localhost systemd[1]: libpod-conmon-62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e.scope: Deactivated successfully. Nov 27 05:05:30 localhost podman[315084]: 2025-11-27 10:05:30.028255506 +0000 UTC m=+0.129631142 container remove 62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:05:30 localhost ovn_controller[156436]: 2025-11-27T10:05:30Z|00146|binding|INFO|Releasing lport 05855691-95b3-4737-b777-8cb539ffdd30 from this chassis (sb_readonly=0) Nov 27 05:05:30 localhost ovn_controller[156436]: 2025-11-27T10:05:30Z|00147|binding|INFO|Setting lport 05855691-95b3-4737-b777-8cb539ffdd30 down in Southbound Nov 27 05:05:30 localhost kernel: device tap05855691-95 left promiscuous mode Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.040 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:30 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:30.048 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=05855691-95b3-4737-b777-8cb539ffdd30) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:30 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:30.051 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 05855691-95b3-4737-b777-8cb539ffdd30 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:30 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:30.052 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:30 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:30.053 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[1b5d4d9f-559c-4fa2-b602-46b0a6395fe2]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.062 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:05:30 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/402514958' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.123 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.476s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.130 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.147 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.150 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.150 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.685s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:05:30 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:30.359 265123 INFO neutron.agent.dhcp.agent [None req-1d1f108b-950a-4cdc-9284-57feae4bbfda - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.461 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:30 localhost nova_compute[284026]: 2025-11-27 10:05:30.779 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e122 e122: 6 total, 6 up, 6 in Nov 27 05:05:30 localhost systemd[1]: var-lib-containers-storage-overlay-83e01c47ecdd532f5a55d83463f2c6b070c25b1662486e56988633d11c0eaff0-merged.mount: Deactivated successfully. Nov 27 05:05:30 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-62d3a610b904534fbb844ad757e9167fb0ecb32c74d506313475d7262016976e-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:30 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:31.134 2 INFO neutron.agent.securitygroups_rpc [None req-c2a49609-ea02-4feb-bba7-eb9d4d416df9 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v210: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 7.1 KiB/s rd, 1.1 KiB/s wr, 10 op/s Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.151 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.151 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.152 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.258 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.259 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.259 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.260 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:05:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:31.279 265123 INFO neutron.agent.linux.ip_lib [None req-df40aea8-c963-422a-bc8a-54879990a426 - - - - - -] Device tap17aa84ab-8f cannot be used as it has no MAC address#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.302 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost kernel: device tap17aa84ab-8f entered promiscuous mode Nov 27 05:05:31 localhost NetworkManager[5971]: [1764237931.3099] manager: (tap17aa84ab-8f): new Generic device (/org/freedesktop/NetworkManager/Devices/29) Nov 27 05:05:31 localhost ovn_controller[156436]: 2025-11-27T10:05:31Z|00148|binding|INFO|Claiming lport 17aa84ab-8ffe-4993-aba1-f1baae398551 for this chassis. Nov 27 05:05:31 localhost ovn_controller[156436]: 2025-11-27T10:05:31Z|00149|binding|INFO|17aa84ab-8ffe-4993-aba1-f1baae398551: Claiming unknown Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.311 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost systemd-udevd[315124]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:31.318 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=17aa84ab-8ffe-4993-aba1-f1baae398551) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:31.320 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 17aa84ab-8ffe-4993-aba1-f1baae398551 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:31.321 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:31.322 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0f51d174-7925-47e1-86ae-695fdce6c1d5]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:31 localhost ovn_controller[156436]: 2025-11-27T10:05:31Z|00150|binding|INFO|Setting lport 17aa84ab-8ffe-4993-aba1-f1baae398551 ovn-installed in OVS Nov 27 05:05:31 localhost ovn_controller[156436]: 2025-11-27T10:05:31Z|00151|binding|INFO|Setting lport 17aa84ab-8ffe-4993-aba1-f1baae398551 up in Southbound Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.342 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost journal[232028]: ethtool ioctl error on tap17aa84ab-8f: No such device Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.377 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost nova_compute[284026]: 2025-11-27 10:05:31.411 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:31.684 2 INFO neutron.agent.securitygroups_rpc [req-c01b8da9-af05-4d94-aae5-9ee02dfd6151 req-cdfb775e-eb0f-4b7c-9231-648ab3b79015 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['836e0029-e4c2-4b3f-8c63-a6373a4acb28']#033[00m Nov 27 05:05:32 localhost nova_compute[284026]: 2025-11-27 10:05:32.143 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:05:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:32.166 2 INFO neutron.agent.securitygroups_rpc [None req-58773de8-af89-4540-9284-fae18ae98907 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:32 localhost nova_compute[284026]: 2025-11-27 10:05:32.183 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:05:32 localhost nova_compute[284026]: 2025-11-27 10:05:32.184 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:05:32 localhost podman[315195]: Nov 27 05:05:32 localhost podman[315195]: 2025-11-27 10:05:32.202995525 +0000 UTC m=+0.096214288 container create 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:32 localhost systemd[1]: Started libpod-conmon-1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1.scope. Nov 27 05:05:32 localhost systemd[1]: tmp-crun.cr1qdc.mount: Deactivated successfully. Nov 27 05:05:32 localhost podman[315195]: 2025-11-27 10:05:32.160135863 +0000 UTC m=+0.053354666 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:32 localhost systemd[1]: Started libcrun container. Nov 27 05:05:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/22b434db09a1373da1b02e2b8d29c0e7dd4b384fb13a5e719db5358a4714abac/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:32 localhost podman[315195]: 2025-11-27 10:05:32.282245501 +0000 UTC m=+0.175464274 container init 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:05:32 localhost podman[315195]: 2025-11-27 10:05:32.347985401 +0000 UTC m=+0.241204154 container start 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:32 localhost dnsmasq[315214]: started, version 2.85 cachesize 150 Nov 27 05:05:32 localhost dnsmasq[315214]: DNS service limited to local subnets Nov 27 05:05:32 localhost dnsmasq[315214]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:32 localhost dnsmasq[315214]: warning: no upstream servers configured Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:05:32 localhost dnsmasq[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e122 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:32 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:32.395 265123 INFO neutron.agent.dhcp.agent [None req-df40aea8-c963-422a-bc8a-54879990a426 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:30Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e9f55cfb-271d-4ff1-a332-23e747d1862f, ip_allocation=immediate, mac_address=fa:16:3e:25:e2:77, name=tempest-NetworksTestDHCPv6-1600819420, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=8, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['69ff9e32-e94d-4f48-9f4f-13e557a87039'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:30Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1110, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:30Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:32 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:32.515 265123 INFO neutron.agent.dhcp.agent [None req-f4359a0d-f76b-4858-b6ff-268040173b98 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:32 localhost dnsmasq[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:32 localhost podman[315233]: 2025-11-27 10:05:32.575792983 +0000 UTC m=+0.051720473 container kill 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:32 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:32.831 265123 INFO neutron.agent.dhcp.agent [None req-2e47e454-e65a-4938-ae58-04337e304634 - - - - - -] DHCP configuration for ports {'e9f55cfb-271d-4ff1-a332-23e747d1862f'} is completed#033[00m Nov 27 05:05:32 localhost dnsmasq[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:32 localhost dnsmasq-dhcp[315214]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:32 localhost podman[315273]: 2025-11-27 10:05:32.901928897 +0000 UTC m=+0.061285121 container kill 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:05:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v211: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 7.1 KiB/s rd, 1.1 KiB/s wr, 10 op/s Nov 27 05:05:33 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:33.416 2 INFO neutron.agent.securitygroups_rpc [req-7a42d185-5711-485a-a446-c99512fa0c04 req-a302c1e8-909a-42d9-a460-696d12bd93b8 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['ba62f992-8c56-4ab7-846b-c35b8592ec3a']#033[00m Nov 27 05:05:33 localhost dnsmasq[315214]: exiting on receipt of SIGTERM Nov 27 05:05:33 localhost podman[315309]: 2025-11-27 10:05:33.466635714 +0000 UTC m=+0.053465289 container kill 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:05:33 localhost systemd[1]: libpod-1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1.scope: Deactivated successfully. Nov 27 05:05:33 localhost podman[315321]: 2025-11-27 10:05:33.539748034 +0000 UTC m=+0.059149653 container died 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:05:33 localhost systemd[1]: tmp-crun.QocSCI.mount: Deactivated successfully. Nov 27 05:05:33 localhost podman[315321]: 2025-11-27 10:05:33.579155251 +0000 UTC m=+0.098556830 container cleanup 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:05:33 localhost systemd[1]: libpod-conmon-1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1.scope: Deactivated successfully. Nov 27 05:05:33 localhost podman[315323]: 2025-11-27 10:05:33.612111314 +0000 UTC m=+0.130409264 container remove 1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:05:33 localhost ovn_controller[156436]: 2025-11-27T10:05:33Z|00152|binding|INFO|Releasing lport 17aa84ab-8ffe-4993-aba1-f1baae398551 from this chassis (sb_readonly=0) Nov 27 05:05:33 localhost ovn_controller[156436]: 2025-11-27T10:05:33Z|00153|binding|INFO|Setting lport 17aa84ab-8ffe-4993-aba1-f1baae398551 down in Southbound Nov 27 05:05:33 localhost kernel: device tap17aa84ab-8f left promiscuous mode Nov 27 05:05:33 localhost nova_compute[284026]: 2025-11-27 10:05:33.656 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:33.666 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=17aa84ab-8ffe-4993-aba1-f1baae398551) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:33.668 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 17aa84ab-8ffe-4993-aba1-f1baae398551 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:33.669 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:33.670 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b7be8812-bf88-46f9-9384-7e30b1817d19]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:33 localhost nova_compute[284026]: 2025-11-27 10:05:33.679 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:33 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:33.689 2 INFO neutron.agent.securitygroups_rpc [req-65b69294-508a-4efe-a575-530ee1d04149 req-b23202b5-47a4-4c01-bc98-48f79b2a2fa5 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['ba62f992-8c56-4ab7-846b-c35b8592ec3a']#033[00m Nov 27 05:05:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:34.151 2 INFO neutron.agent.securitygroups_rpc [req-43e492f5-83e2-426f-b2f4-af80765a25c5 req-14823b0d-29f7-4f81-9add-a82f16fb5466 971a99c977d140c79ac844d1f085ac59 95896ea9103545a7847c0d02928a81fc - - default default] Security group rule updated ['ba62f992-8c56-4ab7-846b-c35b8592ec3a']#033[00m Nov 27 05:05:34 localhost systemd[1]: var-lib-containers-storage-overlay-22b434db09a1373da1b02e2b8d29c0e7dd4b384fb13a5e719db5358a4714abac-merged.mount: Deactivated successfully. Nov 27 05:05:34 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-1334c78fc3c7131279a6ab4af834a654d417a09d9e2e11517aaab146dc9d1eb1-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:34 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:34.376 2 INFO neutron.agent.securitygroups_rpc [None req-ba2fc4f6-b6d9-4981-9ad4-4943234e1e24 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.547 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:34Z, description=, device_id=fcde3155-88c5-479e-a1d8-1ab8c43b2bbe, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e73c2f68-78c0-480f-a115-6f9cae2d35fd, ip_allocation=immediate, mac_address=fa:16:3e:09:51:c2, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1136, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:05:34Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.760 265123 INFO neutron.agent.linux.ip_lib [None req-ee75f776-42e9-4aff-9ee6-370aa0cb889d - - - - - -] Device tape0a53a4b-a2 cannot be used as it has no MAC address#033[00m Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.788 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost kernel: device tape0a53a4b-a2 entered promiscuous mode Nov 27 05:05:34 localhost NetworkManager[5971]: [1764237934.7960] manager: (tape0a53a4b-a2): new Generic device (/org/freedesktop/NetworkManager/Devices/30) Nov 27 05:05:34 localhost ovn_controller[156436]: 2025-11-27T10:05:34Z|00154|binding|INFO|Claiming lport e0a53a4b-a291-426f-9e87-7c71cc23338a for this chassis. Nov 27 05:05:34 localhost ovn_controller[156436]: 2025-11-27T10:05:34Z|00155|binding|INFO|e0a53a4b-a291-426f-9e87-7c71cc23338a: Claiming unknown Nov 27 05:05:34 localhost systemd-udevd[315387]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.800 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:34.806 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e0a53a4b-a291-426f-9e87-7c71cc23338a) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:34.807 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e0a53a4b-a291-426f-9e87-7c71cc23338a in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:34 localhost ovn_controller[156436]: 2025-11-27T10:05:34Z|00156|binding|INFO|Setting lport e0a53a4b-a291-426f-9e87-7c71cc23338a up in Southbound Nov 27 05:05:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:34.809 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:34 localhost ovn_controller[156436]: 2025-11-27T10:05:34Z|00157|binding|INFO|Setting lport e0a53a4b-a291-426f-9e87-7c71cc23338a ovn-installed in OVS Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.809 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:34.810 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[cc5bc4f7-d431-432c-bd2d-187af64516c1]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.810 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.817 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost systemd[1]: tmp-crun.FLIrdo.mount: Deactivated successfully. Nov 27 05:05:34 localhost dnsmasq[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:05:34 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:34 localhost podman[315370]: 2025-11-27 10:05:34.825255745 +0000 UTC m=+0.086488854 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:34 localhost dnsmasq-dhcp[265450]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.833 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent [None req-52a65e32-e720-40e4-ade7-14ccdeb7ec8d - - - - - -] Unable to reload_allocations dhcp for 38b54160-bf0a-4463-b656-cfe2a3341789.: neutron.privileged.agent.linux.ip_lib.NetworkInterfaceNotFound: Network interface tapb7c183b8-b9 not found in namespace qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789. Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent Traceback (most recent call last): Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/dhcp/agent.py", line 264, in _call_driver Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent rv = getattr(driver, action)(**action_kwargs) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 673, in reload_allocations Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent self.device_manager.update(self.network, self.interface_name) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1899, in update Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent self._set_default_route(network, device_name) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1610, in _set_default_route Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent self._set_default_route_ip_version(network, device_name, Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1539, in _set_default_route_ip_version Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent gateway = device.route.get_gateway(ip_version=ip_version) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 671, in get_gateway Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent routes = self.list_routes(ip_version, scope=scope, table=table) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 656, in list_routes Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent return list_ip_routes(self._parent.namespace, ip_version, scope=scope, Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 1611, in list_ip_routes Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent routes = privileged.list_ip_routes(namespace, ip_version, device=device, Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 333, in wrapped_f Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent return self(f, *args, **kw) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 423, in __call__ Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent do = self.iter(retry_state=retry_state) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 360, in iter Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent return fut.result() Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib64/python3.9/concurrent/futures/_base.py", line 439, in result Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent return self.__get_result() Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib64/python3.9/concurrent/futures/_base.py", line 391, in __get_result Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent raise self._exception Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 426, in __call__ Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent result = fn(*args, **kwargs) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/oslo_privsep/priv_context.py", line 271, in _wrap Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent return self.channel.remote_call(name, args, kwargs, Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/oslo_privsep/daemon.py", line 215, in remote_call Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent raise exc_type(*result[2]) Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent neutron.privileged.agent.linux.ip_lib.NetworkInterfaceNotFound: Network interface tapb7c183b8-b9 not found in namespace qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789. Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.855 265123 ERROR neutron.agent.dhcp.agent #033[00m Nov 27 05:05:34 localhost journal[232028]: ethtool ioctl error on tape0a53a4b-a2: No such device Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.873 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost nova_compute[284026]: 2025-11-27 10:05:34.895 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:34.992 265123 INFO neutron.agent.dhcp.agent [None req-9d80365e-439f-4628-9294-200f007accee - - - - - -] DHCP configuration for ports {'e73c2f68-78c0-480f-a115-6f9cae2d35fd'} is completed#033[00m Nov 27 05:05:35 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:35.026 2 INFO neutron.agent.securitygroups_rpc [None req-2c9d9d36-04a3-47c1-80cb-73e0b889a1fc 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v212: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 7.1 KiB/s rd, 1.1 KiB/s wr, 10 op/s Nov 27 05:05:35 localhost nova_compute[284026]: 2025-11-27 10:05:35.463 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:35 localhost podman[315460]: Nov 27 05:05:35 localhost podman[315460]: 2025-11-27 10:05:35.670384407 +0000 UTC m=+0.102461876 container create 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:35 localhost systemd[1]: Started libpod-conmon-254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39.scope. Nov 27 05:05:35 localhost podman[315460]: 2025-11-27 10:05:35.624975427 +0000 UTC m=+0.057052906 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:35 localhost systemd[1]: Started libcrun container. Nov 27 05:05:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/98b7e40c54aaa4cc7b2fb08d56c0fbeac643e899982052ef3e55a4b71e19af9f/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:35 localhost podman[315460]: 2025-11-27 10:05:35.761982808 +0000 UTC m=+0.194060327 container init 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:05:35 localhost podman[315460]: 2025-11-27 10:05:35.771633359 +0000 UTC m=+0.203710828 container start 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:05:35 localhost dnsmasq[315478]: started, version 2.85 cachesize 150 Nov 27 05:05:35 localhost dnsmasq[315478]: DNS service limited to local subnets Nov 27 05:05:35 localhost dnsmasq[315478]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:35 localhost dnsmasq[315478]: warning: no upstream servers configured Nov 27 05:05:35 localhost dnsmasq[315478]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:35 localhost nova_compute[284026]: 2025-11-27 10:05:35.821 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:35 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:35.833 265123 INFO neutron.agent.dhcp.agent [-] Synchronizing state#033[00m Nov 27 05:05:35 localhost nova_compute[284026]: 2025-11-27 10:05:35.937 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:35 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:35.964 265123 INFO neutron.agent.dhcp.agent [None req-30a0e012-a754-43d7-bc52-cf800a2d9813 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:36.036 265123 INFO neutron.agent.dhcp.agent [None req-eb93d195-f263-478f-9262-af3d0bf44f79 - - - - - -] All active networks have been fetched through RPC.#033[00m Nov 27 05:05:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:36.036 265123 INFO neutron.agent.dhcp.agent [-] Starting network 38b54160-bf0a-4463-b656-cfe2a3341789 dhcp configuration#033[00m Nov 27 05:05:36 localhost dnsmasq[265450]: exiting on receipt of SIGTERM Nov 27 05:05:36 localhost podman[315495]: 2025-11-27 10:05:36.210740594 +0000 UTC m=+0.062812362 container kill aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:36 localhost systemd[1]: libpod-aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b.scope: Deactivated successfully. Nov 27 05:05:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:05:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:05:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e123 e123: 6 total, 6 up, 6 in Nov 27 05:05:36 localhost podman[315512]: 2025-11-27 10:05:36.317309911 +0000 UTC m=+0.073330498 container died aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:05:36 localhost podman[315515]: 2025-11-27 10:05:36.398680185 +0000 UTC m=+0.151813043 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible) Nov 27 05:05:36 localhost podman[315553]: 2025-11-27 10:05:36.445573485 +0000 UTC m=+0.085740954 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vendor=Red Hat, Inc., managed_by=edpm_ansible, release=1755695350, maintainer=Red Hat, Inc., name=ubi9-minimal, io.openshift.expose-services=, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, vcs-type=git, io.buildah.version=1.33.7, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, architecture=x86_64, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 05:05:36 localhost podman[315516]: 2025-11-27 10:05:36.495816846 +0000 UTC m=+0.247495535 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:05:36 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:36.496 2 INFO neutron.agent.securitygroups_rpc [None req-bd2e7ade-2060-4809-9aef-086b5734fc8c 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:36 localhost podman[315516]: 2025-11-27 10:05:36.509341393 +0000 UTC m=+0.261020102 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:05:36 localhost podman[315512]: 2025-11-27 10:05:36.519491937 +0000 UTC m=+0.275512474 container remove aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:36 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:05:36 localhost systemd[1]: libpod-conmon-aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b.scope: Deactivated successfully. Nov 27 05:05:36 localhost podman[315515]: 2025-11-27 10:05:36.56609712 +0000 UTC m=+0.319229958 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:36.577 265123 INFO neutron.agent.linux.ip_lib [-] Device tapb7c183b8-b9 cannot be used as it has no MAC address#033[00m Nov 27 05:05:36 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.603 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:36 localhost kernel: device tapb7c183b8-b9 entered promiscuous mode Nov 27 05:05:36 localhost NetworkManager[5971]: [1764237936.6109] manager: (tapb7c183b8-b9): new Generic device (/org/freedesktop/NetworkManager/Devices/31) Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.611 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:36 localhost ovn_controller[156436]: 2025-11-27T10:05:36Z|00158|binding|INFO|Claiming lport b7c183b8-b917-4950-913d-4efc5e8c5146 for this chassis. Nov 27 05:05:36 localhost ovn_controller[156436]: 2025-11-27T10:05:36Z|00159|binding|INFO|b7c183b8-b917-4950-913d-4efc5e8c5146: Claiming unknown Nov 27 05:05:36 localhost ovn_controller[156436]: 2025-11-27T10:05:36Z|00160|binding|INFO|Setting lport b7c183b8-b917-4950-913d-4efc5e8c5146 ovn-installed in OVS Nov 27 05:05:36 localhost podman[315553]: 2025-11-27 10:05:36.623020231 +0000 UTC m=+0.263187730 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, managed_by=edpm_ansible, vcs-type=git, release=1755695350, io.buildah.version=1.33.7, architecture=x86_64, container_name=openstack_network_exporter, io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., vendor=Red Hat, Inc., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.625 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:36 localhost ovn_controller[156436]: 2025-11-27T10:05:36Z|00161|binding|INFO|Setting lport b7c183b8-b917-4950-913d-4efc5e8c5146 up in Southbound Nov 27 05:05:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:36.638 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '192.168.122.172/24', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-38b54160-bf0a-4463-b656-cfe2a3341789', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-38b54160-bf0a-4463-b656-cfe2a3341789', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '02a308de13ca4518beaddd402ba93936', 'neutron:revision_number': '4', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=4e3b70e4-6958-41d7-8c0c-f71821cf7e22, chassis=[], tunnel_key=4, gateway_chassis=[], requested_chassis=[], logical_port=b7c183b8-b917-4950-913d-4efc5e8c5146) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:36.639 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b7c183b8-b917-4950-913d-4efc5e8c5146 in datapath 38b54160-bf0a-4463-b656-cfe2a3341789 bound to our chassis#033[00m Nov 27 05:05:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:36.642 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 8ecc00d5-87bc-4724-a094-f9c282e748b7 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:05:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:36.642 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 38b54160-bf0a-4463-b656-cfe2a3341789, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:05:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:36.643 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0e59b035-2337-47bd-a3de-b0c0761b13f0]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.648 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost systemd[1]: var-lib-containers-storage-overlay-820854a9a613d02ea934359eb4c7e5166b0371e47f95c2852ae595b265fedb5e-merged.mount: Deactivated successfully. Nov 27 05:05:36 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-aac5f6238018fe8c9e2bb5191006b2523aa49f7f5ebbe4b1053a70fe000f7f7b-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:36 localhost journal[232028]: ethtool ioctl error on tapb7c183b8-b9: No such device Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.689 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:36 localhost nova_compute[284026]: 2025-11-27 10:05:36.717 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v214: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 2.7 KiB/s wr, 29 op/s Nov 27 05:05:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e123 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:37 localhost podman[315670]: Nov 27 05:05:37 localhost podman[315670]: 2025-11-27 10:05:37.479263785 +0000 UTC m=+0.074302443 container create 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:05:37 localhost systemd[1]: Started libpod-conmon-8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3.scope. Nov 27 05:05:37 localhost systemd[1]: Started libcrun container. Nov 27 05:05:37 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4e5af62660f5521b5f3caf21e54f5a0a104936140457f420e61fab0585ebd128/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:37 localhost podman[315670]: 2025-11-27 10:05:37.439906659 +0000 UTC m=+0.034945367 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:37 localhost podman[315670]: 2025-11-27 10:05:37.546328751 +0000 UTC m=+0.141367429 container init 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125) Nov 27 05:05:37 localhost podman[315670]: 2025-11-27 10:05:37.552314243 +0000 UTC m=+0.147352931 container start 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:05:37 localhost dnsmasq[315688]: started, version 2.85 cachesize 150 Nov 27 05:05:37 localhost dnsmasq[315688]: DNS service limited to local subnets Nov 27 05:05:37 localhost dnsmasq[315688]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:37 localhost dnsmasq[315688]: warning: no upstream servers configured Nov 27 05:05:37 localhost dnsmasq-dhcp[315688]: DHCP, static leases only on 192.168.122.0, lease time 1d Nov 27 05:05:37 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:05:37 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:37 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:37 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:37.603 265123 INFO neutron.agent.dhcp.agent [None req-8ff7c290-8afb-4eca-b210-124fdba3303e - - - - - -] Finished network 38b54160-bf0a-4463-b656-cfe2a3341789 dhcp configuration#033[00m Nov 27 05:05:37 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:37.604 265123 INFO neutron.agent.dhcp.agent [None req-eb93d195-f263-478f-9262-af3d0bf44f79 - - - - - -] Synchronizing state complete#033[00m Nov 27 05:05:37 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:37.924 2 INFO neutron.agent.securitygroups_rpc [None req-2f5424a0-8d87-4bf0-86c4-8aefc61f42bb 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:37 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:37.947 265123 INFO neutron.agent.dhcp.agent [None req-fbe27904-d801-41d6-a772-a7e2ef707227 - - - - - -] DHCP configuration for ports {'e73c2f68-78c0-480f-a115-6f9cae2d35fd', 'ae3c632e-2fdb-4126-8400-d67f8e72b3a7', '4c6073cf-7b59-4dc4-b067-aa2752dc8c42', '021fcb61-e46f-44f9-9333-752b9ab271fd', 'f5e9ec78-6a6e-47dc-8e13-23fc2324166e', 'b7c183b8-b917-4950-913d-4efc5e8c5146'} is completed#033[00m Nov 27 05:05:37 localhost dnsmasq[315478]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:37 localhost podman[315707]: 2025-11-27 10:05:37.96754618 +0000 UTC m=+0.060037736 container kill 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:05:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:38.116 265123 INFO neutron.agent.dhcp.agent [None req-47c1a754-23aa-4a0a-8180-74acbc0c5b0e - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:34Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=dfe648e0-aa6b-4f5f-aa78-02c662847f10, ip_allocation=immediate, mac_address=fa:16:3e:1e:38:da, name=tempest-NetworksTestDHCPv6-500725626, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=10, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['0d42ce75-90bb-4d05-a726-9377f1959ac6'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:33Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1133, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:34Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:38.277 265123 INFO neutron.agent.dhcp.agent [None req-ba56ab59-1245-4d55-82d8-b862da3215d7 - - - - - -] DHCP configuration for ports {'e0a53a4b-a291-426f-9e87-7c71cc23338a', '3d7f9da0-f1b3-4530-a320-7aaf72a6c2e9', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:38 localhost dnsmasq[315478]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:05:38 localhost podman[315745]: 2025-11-27 10:05:38.326967237 +0000 UTC m=+0.076821863 container kill 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:05:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:38.455 265123 INFO neutron.agent.dhcp.agent [None req-47c1a754-23aa-4a0a-8180-74acbc0c5b0e - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:36Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3d7f9da0-f1b3-4530-a320-7aaf72a6c2e9, ip_allocation=immediate, mac_address=fa:16:3e:22:56:9c, name=tempest-NetworksTestDHCPv6-724649767, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=12, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['d0113f7e-d472-434e-84d5-2ebe74be1de8'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:35Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1144, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:36Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:38.509 265123 INFO neutron.agent.dhcp.agent [None req-775835fd-64d8-49ff-9e31-c8e9f767ad89 - - - - - -] DHCP configuration for ports {'dfe648e0-aa6b-4f5f-aa78-02c662847f10'} is completed#033[00m Nov 27 05:05:38 localhost podman[315783]: 2025-11-27 10:05:38.643863411 +0000 UTC m=+0.059168274 container kill 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:38 localhost dnsmasq[315478]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:05:38 localhost podman[242678]: time="2025-11-27T10:05:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:05:38 localhost podman[242678]: @ - - [27/Nov/2025:10:05:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159231 "" "Go-http-client/1.1" Nov 27 05:05:38 localhost podman[242678]: @ - - [27/Nov/2025:10:05:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20204 "" "Go-http-client/1.1" Nov 27 05:05:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:38.835 265123 INFO neutron.agent.dhcp.agent [None req-1d1c2c12-35e3-40f8-a650-954f2400c26a - - - - - -] DHCP configuration for ports {'3d7f9da0-f1b3-4530-a320-7aaf72a6c2e9'} is completed#033[00m Nov 27 05:05:38 localhost dnsmasq[315478]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:38 localhost podman[315823]: 2025-11-27 10:05:38.992669118 +0000 UTC m=+0.049808159 container kill 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:05:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v215: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 1.3 KiB/s wr, 16 op/s Nov 27 05:05:39 localhost dnsmasq[315478]: exiting on receipt of SIGTERM Nov 27 05:05:39 localhost podman[315861]: 2025-11-27 10:05:39.385624763 +0000 UTC m=+0.062249168 container kill 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:05:39 localhost systemd[1]: libpod-254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39.scope: Deactivated successfully. Nov 27 05:05:39 localhost podman[315876]: 2025-11-27 10:05:39.455454364 +0000 UTC m=+0.059068591 container died 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:39 localhost nova_compute[284026]: 2025-11-27 10:05:39.466 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:39 localhost podman[315876]: 2025-11-27 10:05:39.485804376 +0000 UTC m=+0.089418563 container cleanup 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:05:39 localhost systemd[1]: libpod-conmon-254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39.scope: Deactivated successfully. Nov 27 05:05:39 localhost podman[315878]: 2025-11-27 10:05:39.539254454 +0000 UTC m=+0.132538221 container remove 254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:39 localhost nova_compute[284026]: 2025-11-27 10:05:39.550 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:39 localhost ovn_controller[156436]: 2025-11-27T10:05:39Z|00162|binding|INFO|Releasing lport e0a53a4b-a291-426f-9e87-7c71cc23338a from this chassis (sb_readonly=0) Nov 27 05:05:39 localhost ovn_controller[156436]: 2025-11-27T10:05:39Z|00163|binding|INFO|Setting lport e0a53a4b-a291-426f-9e87-7c71cc23338a down in Southbound Nov 27 05:05:39 localhost kernel: device tape0a53a4b-a2 left promiscuous mode Nov 27 05:05:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:39.560 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e0a53a4b-a291-426f-9e87-7c71cc23338a) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:39.563 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e0a53a4b-a291-426f-9e87-7c71cc23338a in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:39.566 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:39.567 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[972ddff9-2ebd-4f0d-ba63-3638064df20f]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:39 localhost nova_compute[284026]: 2025-11-27 10:05:39.569 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:39 localhost systemd[1]: var-lib-containers-storage-overlay-98b7e40c54aaa4cc7b2fb08d56c0fbeac643e899982052ef3e55a4b71e19af9f-merged.mount: Deactivated successfully. Nov 27 05:05:39 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-254ddda420d6216511958745213cc1ab6b7d5a3e7f57d5c4093857cfffbbdd39-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:39 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:40.173 2 INFO neutron.agent.securitygroups_rpc [None req-e7081121-b355-4e12-9a56-702714a059e3 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.466 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:40.556 265123 INFO neutron.agent.linux.ip_lib [None req-675cc38f-a81b-4f5e-90a1-4b15e7989831 - - - - - -] Device tapccc1ce77-c0 cannot be used as it has no MAC address#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.578 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost kernel: device tapccc1ce77-c0 entered promiscuous mode Nov 27 05:05:40 localhost NetworkManager[5971]: [1764237940.5868] manager: (tapccc1ce77-c0): new Generic device (/org/freedesktop/NetworkManager/Devices/32) Nov 27 05:05:40 localhost ovn_controller[156436]: 2025-11-27T10:05:40Z|00164|binding|INFO|Claiming lport ccc1ce77-c087-4976-ab4a-c4704a716a1d for this chassis. Nov 27 05:05:40 localhost ovn_controller[156436]: 2025-11-27T10:05:40Z|00165|binding|INFO|ccc1ce77-c087-4976-ab4a-c4704a716a1d: Claiming unknown Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.589 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost systemd-udevd[315916]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:40 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:40.600 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=ccc1ce77-c087-4976-ab4a-c4704a716a1d) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:40 localhost ovn_controller[156436]: 2025-11-27T10:05:40Z|00166|binding|INFO|Setting lport ccc1ce77-c087-4976-ab4a-c4704a716a1d ovn-installed in OVS Nov 27 05:05:40 localhost ovn_controller[156436]: 2025-11-27T10:05:40Z|00167|binding|INFO|Setting lport ccc1ce77-c087-4976-ab4a-c4704a716a1d up in Southbound Nov 27 05:05:40 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:40.602 162092 INFO neutron.agent.ovn.metadata.agent [-] Port ccc1ce77-c087-4976-ab4a-c4704a716a1d in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:40 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:40.604 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.604 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:40.605 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[f6bb7cd5-5f01-4973-8e7e-6b796e96935b]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.610 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.634 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost journal[232028]: ethtool ioctl error on tapccc1ce77-c0: No such device Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.672 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.699 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:40 localhost nova_compute[284026]: 2025-11-27 10:05:40.823 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v216: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 1.9 KiB/s wr, 29 op/s Nov 27 05:05:41 localhost podman[315987]: Nov 27 05:05:41 localhost podman[315987]: 2025-11-27 10:05:41.553141775 +0000 UTC m=+0.089727522 container create d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:05:41 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:41.577 2 INFO neutron.agent.securitygroups_rpc [None req-f09b8663-e71f-4244-a4c2-79218c83aa77 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:05:41 localhost systemd[1]: Started libpod-conmon-d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc.scope. Nov 27 05:05:41 localhost podman[315987]: 2025-11-27 10:05:41.510304455 +0000 UTC m=+0.046890212 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:41 localhost systemd[1]: Started libcrun container. Nov 27 05:05:41 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c5b5a85b6e4abb407ecb3061c67960de8cd400ec06f31646d819671a8cbbeb52/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:41 localhost podman[315987]: 2025-11-27 10:05:41.625428713 +0000 UTC m=+0.162014450 container init d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:05:41 localhost podman[315987]: 2025-11-27 10:05:41.635489616 +0000 UTC m=+0.172075353 container start d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:05:41 localhost dnsmasq[316005]: started, version 2.85 cachesize 150 Nov 27 05:05:41 localhost dnsmasq[316005]: DNS service limited to local subnets Nov 27 05:05:41 localhost dnsmasq[316005]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:41 localhost dnsmasq[316005]: warning: no upstream servers configured Nov 27 05:05:41 localhost dnsmasq-dhcp[316005]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:05:41 localhost dnsmasq[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:41 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:41 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:41 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:41.691 265123 INFO neutron.agent.dhcp.agent [None req-675cc38f-a81b-4f5e-90a1-4b15e7989831 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:39Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=1b9d2f5f-18be-470c-a868-ba80dca93eea, ip_allocation=immediate, mac_address=fa:16:3e:2d:63:7e, name=tempest-NetworksTestDHCPv6-1895964485, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=14, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['c5692104-32da-43b7-b543-bf474b2b30da'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:39Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1177, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:39Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:41 localhost dnsmasq[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:41 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:41 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:41 localhost podman[316024]: 2025-11-27 10:05:41.861068436 +0000 UTC m=+0.045765741 container kill d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:41 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:41.892 265123 INFO neutron.agent.dhcp.agent [None req-14aed376-7f44-4d7c-89f4-36f6b20b2edd - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:42 localhost ovn_controller[156436]: 2025-11-27T10:05:42Z|00168|binding|INFO|Releasing lport ccc1ce77-c087-4976-ab4a-c4704a716a1d from this chassis (sb_readonly=0) Nov 27 05:05:42 localhost ovn_controller[156436]: 2025-11-27T10:05:42Z|00169|binding|INFO|Setting lport ccc1ce77-c087-4976-ab4a-c4704a716a1d down in Southbound Nov 27 05:05:42 localhost nova_compute[284026]: 2025-11-27 10:05:42.035 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:42 localhost kernel: device tapccc1ce77-c0 left promiscuous mode Nov 27 05:05:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:42.046 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=ccc1ce77-c087-4976-ab4a-c4704a716a1d) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:42.049 162092 INFO neutron.agent.ovn.metadata.agent [-] Port ccc1ce77-c087-4976-ab4a-c4704a716a1d in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:42.050 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:42.051 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[6ee6707c-45d5-4288-8b6b-9c0eca022984]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:42 localhost nova_compute[284026]: 2025-11-27 10:05:42.065 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.111 265123 INFO neutron.agent.dhcp.agent [None req-ba1bec23-226c-4118-8247-36569df75e5b - - - - - -] DHCP configuration for ports {'1b9d2f5f-18be-470c-a868-ba80dca93eea'} is completed#033[00m Nov 27 05:05:42 localhost dnsmasq[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:42 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:42 localhost dnsmasq-dhcp[316005]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:42 localhost podman[316063]: 2025-11-27 10:05:42.27394291 +0000 UTC m=+0.061849946 container kill d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent [None req-675cc38f-a81b-4f5e-90a1-4b15e7989831 - - - - - -] Unable to reload_allocations dhcp for 1ba86240-13a7-470b-9c24-36a566ecc5cd.: neutron.privileged.agent.linux.ip_lib.NetworkInterfaceNotFound: Network interface tapccc1ce77-c0 not found in namespace qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd. Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent Traceback (most recent call last): Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/dhcp/agent.py", line 264, in _call_driver Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent rv = getattr(driver, action)(**action_kwargs) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 673, in reload_allocations Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent self.device_manager.update(self.network, self.interface_name) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1899, in update Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent self._set_default_route(network, device_name) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1610, in _set_default_route Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent self._set_default_route_ip_version(network, device_name, Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/dhcp.py", line 1539, in _set_default_route_ip_version Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent gateway = device.route.get_gateway(ip_version=ip_version) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 671, in get_gateway Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent routes = self.list_routes(ip_version, scope=scope, table=table) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 656, in list_routes Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent return list_ip_routes(self._parent.namespace, ip_version, scope=scope, Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/neutron/agent/linux/ip_lib.py", line 1611, in list_ip_routes Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent routes = privileged.list_ip_routes(namespace, ip_version, device=device, Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 333, in wrapped_f Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent return self(f, *args, **kw) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 423, in __call__ Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent do = self.iter(retry_state=retry_state) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 360, in iter Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent return fut.result() Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib64/python3.9/concurrent/futures/_base.py", line 439, in result Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent return self.__get_result() Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib64/python3.9/concurrent/futures/_base.py", line 391, in __get_result Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent raise self._exception Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/tenacity/__init__.py", line 426, in __call__ Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent result = fn(*args, **kwargs) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/oslo_privsep/priv_context.py", line 271, in _wrap Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent return self.channel.remote_call(name, args, kwargs, Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent File "/usr/lib/python3.9/site-packages/oslo_privsep/daemon.py", line 215, in remote_call Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent raise exc_type(*result[2]) Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent neutron.privileged.agent.linux.ip_lib.NetworkInterfaceNotFound: Network interface tapccc1ce77-c0 not found in namespace qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd. Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.299 265123 ERROR neutron.agent.dhcp.agent #033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.302 265123 INFO neutron.agent.dhcp.agent [None req-eb93d195-f263-478f-9262-af3d0bf44f79 - - - - - -] Synchronizing state#033[00m Nov 27 05:05:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e123 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.493 265123 INFO neutron.agent.dhcp.agent [None req-7fcb19df-1928-4d32-ac80-4f52e06fac47 - - - - - -] All active networks have been fetched through RPC.#033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.494 265123 INFO neutron.agent.dhcp.agent [-] Starting network 1ba86240-13a7-470b-9c24-36a566ecc5cd dhcp configuration#033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.494 265123 INFO neutron.agent.dhcp.agent [-] Finished network 1ba86240-13a7-470b-9c24-36a566ecc5cd dhcp configuration#033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.495 265123 INFO neutron.agent.dhcp.agent [None req-7fcb19df-1928-4d32-ac80-4f52e06fac47 - - - - - -] Synchronizing state complete#033[00m Nov 27 05:05:42 localhost systemd[1]: tmp-crun.pWEHCD.mount: Deactivated successfully. Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.642 265123 INFO neutron.agent.dhcp.agent [None req-291ee244-d8c3-467a-a9c8-27fc013c85cb - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:42.732 265123 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'neutron-rootwrap', '/etc/neutron/rootwrap.conf', 'privsep-helper', '--config-file', '/etc/neutron/neutron.conf', '--config-dir', '/etc/neutron.conf.d', '--privsep_context', 'neutron.privileged.dhcp_release_cmd', '--privsep_sock_path', '/tmp/tmp7ea0ukqs/privsep.sock']#033[00m Nov 27 05:05:42 localhost dnsmasq[316005]: exiting on receipt of SIGTERM Nov 27 05:05:42 localhost systemd[1]: tmp-crun.gZkzMS.mount: Deactivated successfully. Nov 27 05:05:42 localhost podman[316095]: 2025-11-27 10:05:42.785887447 +0000 UTC m=+0.070781238 container kill d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:42 localhost systemd[1]: libpod-d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc.scope: Deactivated successfully. Nov 27 05:05:42 localhost podman[316112]: 2025-11-27 10:05:42.860596701 +0000 UTC m=+0.058905097 container died d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:05:42 localhost podman[316112]: 2025-11-27 10:05:42.894967331 +0000 UTC m=+0.093275677 container cleanup d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:05:42 localhost systemd[1]: libpod-conmon-d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc.scope: Deactivated successfully. Nov 27 05:05:42 localhost podman[316114]: 2025-11-27 10:05:42.943725883 +0000 UTC m=+0.131756650 container remove d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:05:43 localhost ovn_controller[156436]: 2025-11-27T10:05:43Z|00170|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:05:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v217: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 1.9 KiB/s wr, 29 op/s Nov 27 05:05:43 localhost nova_compute[284026]: 2025-11-27 10:05:43.150 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:43.214 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=13, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=12) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:43 localhost nova_compute[284026]: 2025-11-27 10:05:43.214 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:43.215 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:05:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:43.375 265123 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap#033[00m Nov 27 05:05:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:43.250 316141 INFO oslo.privsep.daemon [-] privsep daemon starting#033[00m Nov 27 05:05:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:43.255 316141 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0#033[00m Nov 27 05:05:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:43.259 316141 INFO oslo.privsep.daemon [-] privsep process running with capabilities (eff/prm/inh): CAP_NET_ADMIN|CAP_SYS_ADMIN/CAP_NET_ADMIN|CAP_SYS_ADMIN/none#033[00m Nov 27 05:05:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:43.259 316141 INFO oslo.privsep.daemon [-] privsep daemon running as pid 316141#033[00m Nov 27 05:05:43 localhost systemd[1]: var-lib-containers-storage-overlay-c5b5a85b6e4abb407ecb3061c67960de8cd400ec06f31646d819671a8cbbeb52-merged.mount: Deactivated successfully. Nov 27 05:05:43 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-d4fc844daec35fe94580f9bf23bfc4bbf4bbf5dffbeaa8feb593290ffbaef2bc-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:43 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:43.574 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:05:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:43.575 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:05:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:43.576 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:05:43 localhost dnsmasq-dhcp[315688]: DHCPRELEASE(tapb7c183b8-b9) 192.168.122.188 fa:16:3e:c5:e7:be Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.137 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.138 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.171 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.172 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4765fd0b-1fad-448f-a2e3-2c66da8696e1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.139269', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a31f2078-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': 'fc3a948e8bcdd8e639088586edffdb5e76ff3ec79a7d7c9243d440f402639c07'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.139269', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a31f3a04-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '3bf29c8943c0dcdb999c06cb50917378f043478e3adfe13207fbcfd761f9b1fa'}]}, 'timestamp': '2025-11-27 10:05:44.173575', '_unique_id': '7cf1e542d07b43c7b338938cb952488a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.177 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.177 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.181 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8442adcb-7e6a-455a-92db-db91b6996044', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.177760', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32088c8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '8d55a86f96334cc8d642b854c6889dff6033018ec1a6e56209eb3c90ec6764f6'}]}, 'timestamp': '2025-11-27 10:05:44.182170', '_unique_id': '92dc75ced8c747768b250d0dc4e62f4f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.184 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.185 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.185 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'db1601c2-5226-4f5f-93ab-c5a3417413e8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.185215', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a321190a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '329c25144af66e12867a48ed250b864ff1cbadec875b68a28aa17fe12556200f'}]}, 'timestamp': '2025-11-27 10:05:44.185746', '_unique_id': '54be0540352949dcb03624ad57f01ff0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.188 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.201 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.201 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6fd7d4b1-0dc8-4054-9765-79160360e2ac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.188357', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a32385fa-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': 'fc2714eba046c0391b28da022574022eb82507ae07c35e86a437ab82877672e6'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.188357', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a3239f18-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': '29f4264be50eb84ded180f330bdded000a0ee5f27805a71d20c39e67374de300'}]}, 'timestamp': '2025-11-27 10:05:44.202279', '_unique_id': '1757057f13c441f0ada0d7363fa6b42e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.203 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.204 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.205 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '38a90ade-b7bc-49fe-b34d-c73b8e5d4bc4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.205010', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a3241db2-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '8b05938167f5452efa4c27e86eb38644e55bcd3202bdeba5a46087d33fa2c6d0'}]}, 'timestamp': '2025-11-27 10:05:44.205488', '_unique_id': '8c0682ec60d54eb69f9e787422dd56a3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.206 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.207 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.208 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.208 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4374e635-76ea-4a2b-aa5a-f8dc593c0573', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.208041', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a3249396-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': 'cb06bf35d722e6793f4812842feab7314ee139beb2880693d4839b014bd1b7f7'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.208041', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a324a58e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': 'e6e1a8790ed69036f3064fe308e01d98e74e0426d8fd8439d9fcc884b2785286'}]}, 'timestamp': '2025-11-27 10:05:44.208934', '_unique_id': '46c053de0ab4468ca976a302e8a8dbab'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.212 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.212 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.213 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8d055709-6a09-419e-9a35-2753fedc6c05', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.212717', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a3255038-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '02f2becb5f7686ea1deb456b38168584ee4940b8562085750ae9d67751244524'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.212717', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a32574e6-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '737db6025187db4a948e9ab79f1f28cfbcf2519bd18a6cd989f4ed47b82d4980'}]}, 'timestamp': '2025-11-27 10:05:44.214346', '_unique_id': '20aeed98b23148e7b18846709a44ee67'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.217 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.219 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.219 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'db13eceb-7538-45b7-a773-3259ca48cfaa', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.219810', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a32663a6-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '4b18c4d97e0a3e3436873529fffac24db5e7c748778b8d31d2d9f507ed43e41e'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.219810', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a326783c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': 'eba0507db1e931093601c4774c65ad4cd05cdcc56e5906167828af47e869ffa4'}]}, 'timestamp': '2025-11-27 10:05:44.220907', '_unique_id': '6415d1ee35884fe69100167229e4e8d7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.223 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 16090000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '15e77044-93b5-48e3-aa7e-9564d1023c24', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 16090000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:05:44.223646', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'a329e44a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.430481684, 'message_signature': 'd0199ca7058c84076df8a7d861d67b5967fe04ef14e90f4df934bc8b4d28de1d'}]}, 'timestamp': '2025-11-27 10:05:44.243483', '_unique_id': '7a902f4c21a349148de419dced45e159'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ae4629ef-5965-405a-8d7b-bfd4c21508dc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.247440', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a32a9c8c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': 'f5f76d8f1105eb47c4a7808e3fe960e72dd72e20d9703c84b2a8989a14bf4190'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.247440', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a32ab2ee-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '934a907858119ed9f9c856397ed029c23bc09698ad890e64ac0c0d2f3e620895'}]}, 'timestamp': '2025-11-27 10:05:44.248718', '_unique_id': 'a3560b459ee149d89225e5df45440728'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.251 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.251 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.251 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '921aaa30-2120-4451-9a3e-71a70a8ae2bd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.251327', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a32b2fda-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': 'b764a50ff63d629b95a5fbf6f1554c18022e01a95feb166170ca9b22bb0ccbac'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.251327', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a32b40ba-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.376554294, 'message_signature': '6b9670ac31c82fc96c3f92bbe602bf124fa5cc8eeea5f29755928af5608292da'}]}, 'timestamp': '2025-11-27 10:05:44.252229', '_unique_id': '801f0acd6aca469c8cd9c05647e69848'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.254 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.254 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '64442e9a-69a5-4e8f-a6f2-07e8d482ad66', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.254469', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32bab0e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '982dc01b4336dd0d45f999b64efb005564f20a2962097ee96e9ba72194c8685a'}]}, 'timestamp': '2025-11-27 10:05:44.254982', '_unique_id': 'd016923e40e348baacd242c384d73c62'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.257 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.257 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.258 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4247eced-458a-410f-9719-2eb814370bf2', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:05:44.258377', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'a32c43de-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.430481684, 'message_signature': 'd5d33e09f4346cb8b968f22e0980a5d6b4ce1c476aec9f7f0162aa938a61048f'}]}, 'timestamp': '2025-11-27 10:05:44.258880', '_unique_id': '539ad096b5014e2e8eb0019fac72a906'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.261 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f9134ed2-2315-474e-8c07-c518403a4213', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.261136', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a32cad74-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '7f92ba2f5e6927c72481258b4e5f2a9484651649a8320fb775a956a3a888bf74'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.261136', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a32cbf8a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '49ccb63450a6a19aafb472a97df4d8f082250675d4d9da15e801ea1893c0cd89'}]}, 'timestamp': '2025-11-27 10:05:44.262031', '_unique_id': '8249ee3e3e87457fa6e65c8ad53c048f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.263 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.264 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4e2536b5-fac5-4586-9800-e3dd9c8ef922', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.264378', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32d2df8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': 'd2d62ee56b62c3abb537d3e7c63c3140cc33f424f6d8c7940358da440860a0d4'}]}, 'timestamp': '2025-11-27 10:05:44.264886', '_unique_id': 'd34d8a2eaa854b29a35f712f3f2e794b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.275 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.275 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '21074935-b15c-4b6d-9b8c-0bd9294b7415', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.275482', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32ee09e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '64a190c1c95838979f73d1f21644c302a87875dc4f44db4589bb76d69d7edfb0'}]}, 'timestamp': '2025-11-27 10:05:44.276024', '_unique_id': '70affbcc5bf34e219df35bcf656c96ef'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.277 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.278 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.278 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8326c3d8-edb1-461d-8984-6fb440c2212d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.278244', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32f4ae8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '823a0f3c79be6b82998bd42042916c5a4f4b98cd3235216de3e4546b99afa3c4'}]}, 'timestamp': '2025-11-27 10:05:44.278764', '_unique_id': 'ea9f5f3d912f440393eda05b20a421b4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.279 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.280 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.281 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c381cf0c-3829-44a1-bd66-ee169c5f23f8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.280964', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a32fb528-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': 'ff0193419e91ed4f467a5c6a2da6238ad34c9eb44ee4e9d7a76396bce4166ae1'}]}, 'timestamp': '2025-11-27 10:05:44.281451', '_unique_id': 'aca1c19533dd4a3fbd5a0894e14512b1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.282 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.283 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.283 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '05e5e981-f1dd-4471-b116-3a6dc102e1fd', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.283464', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a33013a6-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '2e902c2bf3e6bc10e2bc764eeccc82cc65d2edef654f2d96ea68d1a9c9c1fcd6'}]}, 'timestamp': '2025-11-27 10:05:44.283785', '_unique_id': 'a7b1a34f77554cf8b9a77c627895d665'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.284 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c5e0318f-96aa-4036-9d35-363962edcf3c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:05:44.285098', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'a3305230-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.365834713, 'message_signature': '081e07f3998e037e8298c2da8a03ccf74903f3801743f66eb5ec23a3c09e7dd3'}]}, 'timestamp': '2025-11-27 10:05:44.285384', '_unique_id': '8507d081d75a42e886aea5b145f5a0e5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.285 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.286 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.286 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '32af4d07-1c04-4f12-91be-4f4830af9508', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:05:44.286750', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'a33092ae-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '7002f3261e16fe91e6fc6a2ccc92a181eaa3b3b7efc42627dca9353ac5bb0bd6'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:05:44.286750', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'a3309c7c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12174.32732936, 'message_signature': '19fe4715cddd4652a4120212a638c0051092abfe436fd2e7d85599b2849cab6b'}]}, 'timestamp': '2025-11-27 10:05:44.287268', '_unique_id': '1fc8beb81075410b89eed571c94578a5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:05:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:05:44.287 12 ERROR oslo_messaging.notify.messaging Nov 27 05:05:44 localhost podman[316161]: 2025-11-27 10:05:44.339145451 +0000 UTC m=+0.071279193 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:05:44 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:05:44 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:44 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:05:44 Nov 27 05:05:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:05:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:05:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['.mgr', 'manila_metadata', 'images', 'manila_data', 'backups', 'vms', 'volumes'] Nov 27 05:05:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:05:44 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:44.538 265123 INFO neutron.agent.linux.ip_lib [None req-6edaf172-c4b0-4808-8f77-dbce6db414a1 - - - - - -] Device tapbeb4101c-44 cannot be used as it has no MAC address#033[00m Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.571 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost kernel: device tapbeb4101c-44 entered promiscuous mode Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.577 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost NetworkManager[5971]: [1764237944.5779] manager: (tapbeb4101c-44): new Generic device (/org/freedesktop/NetworkManager/Devices/33) Nov 27 05:05:44 localhost systemd-udevd[316191]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.583 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:05:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:05:44 localhost ovn_controller[156436]: 2025-11-27T10:05:44Z|00171|binding|INFO|Claiming lport beb4101c-445d-4f4e-8517-c39429819f92 for this chassis. Nov 27 05:05:44 localhost ovn_controller[156436]: 2025-11-27T10:05:44Z|00172|binding|INFO|beb4101c-445d-4f4e-8517-c39429819f92: Claiming unknown Nov 27 05:05:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:44.594 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=beb4101c-445d-4f4e-8517-c39429819f92) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:44.596 162092 INFO neutron.agent.ovn.metadata.agent [-] Port beb4101c-445d-4f4e-8517-c39429819f92 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:44 localhost ovn_controller[156436]: 2025-11-27T10:05:44Z|00173|binding|INFO|Setting lport beb4101c-445d-4f4e-8517-c39429819f92 ovn-installed in OVS Nov 27 05:05:44 localhost ovn_controller[156436]: 2025-11-27T10:05:44Z|00174|binding|INFO|Setting lport beb4101c-445d-4f4e-8517-c39429819f92 up in Southbound Nov 27 05:05:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:44.597 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.598 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:44.598 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b20fb749-2c98-45f6-98a0-b1c836b40c64]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.619 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost journal[232028]: ethtool ioctl error on tapbeb4101c-44: No such device Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.655 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:05:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002170138888888889 quantized to 16 (current 16) Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:05:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:05:44 localhost nova_compute[284026]: 2025-11-27 10:05:44.679 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:05:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:05:44 localhost systemd[1]: tmp-crun.euNlOO.mount: Deactivated successfully. Nov 27 05:05:44 localhost podman[316228]: 2025-11-27 10:05:44.982042695 +0000 UTC m=+0.068076365 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, config_id=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:05:45 localhost podman[316228]: 2025-11-27 10:05:45.014926656 +0000 UTC m=+0.100960296 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:45 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:05:45 localhost podman[316230]: 2025-11-27 10:05:45.028103153 +0000 UTC m=+0.113783484 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:05:45 localhost podman[316230]: 2025-11-27 10:05:45.039866391 +0000 UTC m=+0.125546712 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:05:45 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:05:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v218: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 1.9 KiB/s wr, 29 op/s Nov 27 05:05:45 localhost podman[316308]: Nov 27 05:05:45 localhost podman[316308]: 2025-11-27 10:05:45.428097178 +0000 UTC m=+0.086936977 container create 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:45 localhost systemd[1]: Started libpod-conmon-5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736.scope. Nov 27 05:05:45 localhost nova_compute[284026]: 2025-11-27 10:05:45.469 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:45 localhost systemd[1]: Started libcrun container. Nov 27 05:05:45 localhost podman[316308]: 2025-11-27 10:05:45.386501081 +0000 UTC m=+0.045340910 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:05:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d9dea73b2ce7fac2e549d438af1b0e042b18f9d2f39c1f5ac92e957cd9b18eb5/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:05:45 localhost podman[316308]: 2025-11-27 10:05:45.499246365 +0000 UTC m=+0.158086154 container init 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:45 localhost podman[316308]: 2025-11-27 10:05:45.507752435 +0000 UTC m=+0.166592224 container start 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:05:45 localhost dnsmasq[316327]: started, version 2.85 cachesize 150 Nov 27 05:05:45 localhost dnsmasq[316327]: DNS service limited to local subnets Nov 27 05:05:45 localhost dnsmasq[316327]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:05:45 localhost dnsmasq[316327]: warning: no upstream servers configured Nov 27 05:05:45 localhost dnsmasq-dhcp[316327]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:05:45 localhost dnsmasq[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:45 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:45 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:45.713 265123 INFO neutron.agent.dhcp.agent [None req-d6bf852b-4a05-454e-a4b6-03065c1e3ba1 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:05:45 localhost nova_compute[284026]: 2025-11-27 10:05:45.826 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:45 localhost dnsmasq[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:45 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:45 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:45 localhost podman[316345]: 2025-11-27 10:05:45.833021875 +0000 UTC m=+0.065198796 container kill 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:46 localhost dnsmasq-dhcp[315688]: DHCPRELEASE(tapb7c183b8-b9) 192.168.122.180 fa:16:3e:09:51:c2 Nov 27 05:05:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:46.129 265123 INFO neutron.agent.dhcp.agent [None req-a4660050-e3c7-4f65-b07d-5ff39fb21540 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea', 'beb4101c-445d-4f4e-8517-c39429819f92'} is completed#033[00m Nov 27 05:05:46 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:46.217 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '13'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:05:46 localhost ovn_controller[156436]: 2025-11-27T10:05:46Z|00175|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:05:46 localhost nova_compute[284026]: 2025-11-27 10:05:46.440 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:46 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:05:46 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:46 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:46 localhost podman[316387]: 2025-11-27 10:05:46.684859359 +0000 UTC m=+0.060499799 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e124 e124: 6 total, 6 up, 6 in Nov 27 05:05:47 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:47.046 2 INFO neutron.agent.securitygroups_rpc [None req-49c07172-523c-4c24-9c18-3b718b79ea0a 4e42ff97234741689bf0cfbe84b67105 f4460e56db744027b7a29a83ceb92d47 - - default default] Security group member updated ['741444dc-4e14-49f0-8033-d72c008fc62c']#033[00m Nov 27 05:05:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:47.090 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:46Z, description=, device_id=2afdd8f5-61b6-43fc-87ec-f3a1c4222e92, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=be1a819d-7837-4b06-9086-538e06a2c2a9, ip_allocation=immediate, mac_address=fa:16:3e:22:d3:23, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=18, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['06fc6b8e-cc23-4c86-bb45-2bedc5657480'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:44Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=False, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1225, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:46Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v220: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 818 B/s wr, 16 op/s Nov 27 05:05:47 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:47.173 2 INFO neutron.agent.securitygroups_rpc [None req-49c07172-523c-4c24-9c18-3b718b79ea0a 4e42ff97234741689bf0cfbe84b67105 f4460e56db744027b7a29a83ceb92d47 - - default default] Security group member updated ['741444dc-4e14-49f0-8033-d72c008fc62c']#033[00m Nov 27 05:05:47 localhost podman[316426]: 2025-11-27 10:05:47.272686212 +0000 UTC m=+0.059026849 container kill 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:05:47 localhost dnsmasq[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:47 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:47 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e124 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:47.513 265123 INFO neutron.agent.dhcp.agent [None req-016d2d3b-de95-4857-9e2b-898c6604f026 - - - - - -] DHCP configuration for ports {'be1a819d-7837-4b06-9086-538e06a2c2a9'} is completed#033[00m Nov 27 05:05:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e125 e125: 6 total, 6 up, 6 in Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #28. Immutable memtables: 0. Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.743836) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 13] Flushing memtable with next log file: 28 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947743895, "job": 13, "event": "flush_started", "num_memtables": 1, "num_entries": 2602, "num_deletes": 264, "total_data_size": 3725540, "memory_usage": 3786800, "flush_reason": "Manual Compaction"} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 13] Level-0 flush table #29: started Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947762163, "cf_name": "default", "job": 13, "event": "table_file_creation", "file_number": 29, "file_size": 2391894, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 17607, "largest_seqno": 20204, "table_properties": {"data_size": 2382343, "index_size": 5992, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2501, "raw_key_size": 20957, "raw_average_key_size": 21, "raw_value_size": 2362815, "raw_average_value_size": 2408, "num_data_blocks": 260, "num_entries": 981, "num_filter_entries": 981, "num_deletions": 264, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237783, "oldest_key_time": 1764237783, "file_creation_time": 1764237947, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 29, "seqno_to_time_mapping": "N/A"}} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 13] Flush lasted 18370 microseconds, and 5984 cpu microseconds. Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.762208) [db/flush_job.cc:967] [default] [JOB 13] Level-0 flush table #29: 2391894 bytes OK Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.762233) [db/memtable_list.cc:519] [default] Level-0 commit table #29 started Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.765387) [db/memtable_list.cc:722] [default] Level-0 commit table #29: memtable #1 done Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.765412) EVENT_LOG_v1 {"time_micros": 1764237947765406, "job": 13, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.765431) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 13] Try to delete WAL files size 3713890, prev total WAL file size 3713890, number of live WAL files 2. Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000025.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.766429) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003131373937' seq:72057594037927935, type:22 .. '7061786F73003132303439' seq:0, type:0; will stop at (end) Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 14] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 13 Base level 0, inputs: [29(2335KB)], [27(18MB)] Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947766495, "job": 14, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [29], "files_L6": [27], "score": -1, "input_data_size": 21976420, "oldest_snapshot_seqno": -1} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 14] Generated table #30: 12462 keys, 18134888 bytes, temperature: kUnknown Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947907951, "cf_name": "default", "job": 14, "event": "table_file_creation", "file_number": 30, "file_size": 18134888, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 18063681, "index_size": 38978, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 31173, "raw_key_size": 333806, "raw_average_key_size": 26, "raw_value_size": 17851321, "raw_average_value_size": 1432, "num_data_blocks": 1482, "num_entries": 12462, "num_filter_entries": 12462, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764237947, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 30, "seqno_to_time_mapping": "N/A"}} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.908341) [db/compaction/compaction_job.cc:1663] [default] [JOB 14] Compacted 1@0 + 1@6 files to L6 => 18134888 bytes Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.910095) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 155.2 rd, 128.1 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(2.3, 18.7 +0.0 blob) out(17.3 +0.0 blob), read-write-amplify(16.8) write-amplify(7.6) OK, records in: 12999, records dropped: 537 output_compression: NoCompression Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.910124) EVENT_LOG_v1 {"time_micros": 1764237947910111, "job": 14, "event": "compaction_finished", "compaction_time_micros": 141615, "compaction_time_cpu_micros": 46556, "output_level": 6, "num_output_files": 1, "total_output_size": 18134888, "num_input_records": 12999, "num_output_records": 12462, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000029.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947910596, "job": 14, "event": "table_file_deletion", "file_number": 29} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000027.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764237947913310, "job": 14, "event": "table_file_deletion", "file_number": 27} Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.766275) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.913418) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.913424) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.913427) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.913430) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:05:47.913432) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:05:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:47.972 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:47Z, description=, device_id=79f96a39-1b24-48b0-8a93-37e5881b9129, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=1d7742c5-5ea2-4b7e-a42b-cfb3177b57eb, ip_allocation=immediate, mac_address=fa:16:3e:99:0c:06, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1236, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:05:47Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:05:48 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:05:48 localhost podman[316464]: 2025-11-27 10:05:48.16417384 +0000 UTC m=+0.046402867 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:05:48 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:05:48 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:05:48 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:48.191 2 INFO neutron.agent.securitygroups_rpc [None req-ed02356c-af44-4c9b-9637-864320cd3093 4e42ff97234741689bf0cfbe84b67105 f4460e56db744027b7a29a83ceb92d47 - - default default] Security group member updated ['741444dc-4e14-49f0-8033-d72c008fc62c']#033[00m Nov 27 05:05:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:48.350 265123 INFO neutron.agent.dhcp.agent [None req-7592d5f0-034f-4031-830b-2169a1684e0a - - - - - -] DHCP configuration for ports {'1d7742c5-5ea2-4b7e-a42b-cfb3177b57eb'} is completed#033[00m Nov 27 05:05:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:48.619 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:46Z, description=, device_id=2afdd8f5-61b6-43fc-87ec-f3a1c4222e92, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=be1a819d-7837-4b06-9086-538e06a2c2a9, ip_allocation=immediate, mac_address=fa:16:3e:22:d3:23, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=18, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['06fc6b8e-cc23-4c86-bb45-2bedc5657480'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:44Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=False, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1225, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:46Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:05:48 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:48.630 2 INFO neutron.agent.securitygroups_rpc [None req-3283a102-97c4-4c79-aa9c-c50e8a23b992 4e42ff97234741689bf0cfbe84b67105 f4460e56db744027b7a29a83ceb92d47 - - default default] Security group member updated ['741444dc-4e14-49f0-8033-d72c008fc62c']#033[00m Nov 27 05:05:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e126 e126: 6 total, 6 up, 6 in Nov 27 05:05:48 localhost podman[316503]: 2025-11-27 10:05:48.83166192 +0000 UTC m=+0.087428369 container kill 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:05:48 localhost dnsmasq[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:05:48 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:48 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:05:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2959823449' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:05:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:05:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2959823449' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:05:49 localhost nova_compute[284026]: 2025-11-27 10:05:49.105 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:49 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:49.118 265123 INFO neutron.agent.dhcp.agent [None req-0ea5f6ff-4e51-4410-a4ba-2483d59a81a9 - - - - - -] DHCP configuration for ports {'be1a819d-7837-4b06-9086-538e06a2c2a9'} is completed#033[00m Nov 27 05:05:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v223: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail Nov 27 05:05:49 localhost nova_compute[284026]: 2025-11-27 10:05:49.195 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:50 localhost nova_compute[284026]: 2025-11-27 10:05:50.473 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e127 e127: 6 total, 6 up, 6 in Nov 27 05:05:50 localhost nova_compute[284026]: 2025-11-27 10:05:50.829 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v225: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 6.1 KiB/s rd, 1.3 KiB/s wr, 10 op/s Nov 27 05:05:52 localhost podman[316542]: 2025-11-27 10:05:52.070246256 +0000 UTC m=+0.062038543 container kill 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:52 localhost dnsmasq[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:05:52 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:05:52 localhost dnsmasq-dhcp[316327]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:05:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e127 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v226: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 4.5 KiB/s rd, 1023 B/s wr, 7 op/s Nov 27 05:05:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:05:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2816586628' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:05:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:05:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2816586628' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:05:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v227: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 3.6 KiB/s rd, 827 B/s wr, 6 op/s Nov 27 05:05:55 localhost podman[316582]: 2025-11-27 10:05:55.275043866 +0000 UTC m=+0.061359253 container kill 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:05:55 localhost dnsmasq[316327]: exiting on receipt of SIGTERM Nov 27 05:05:55 localhost systemd[1]: libpod-5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736.scope: Deactivated successfully. Nov 27 05:05:55 localhost podman[316595]: 2025-11-27 10:05:55.350859079 +0000 UTC m=+0.062990637 container died 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:05:55 localhost podman[316595]: 2025-11-27 10:05:55.386488714 +0000 UTC m=+0.098620202 container cleanup 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:05:55 localhost systemd[1]: libpod-conmon-5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736.scope: Deactivated successfully. Nov 27 05:05:55 localhost podman[316602]: 2025-11-27 10:05:55.435983565 +0000 UTC m=+0.134197436 container remove 5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:55 localhost nova_compute[284026]: 2025-11-27 10:05:55.450 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:55 localhost kernel: device tapbeb4101c-44 left promiscuous mode Nov 27 05:05:55 localhost ovn_controller[156436]: 2025-11-27T10:05:55Z|00176|binding|INFO|Releasing lport beb4101c-445d-4f4e-8517-c39429819f92 from this chassis (sb_readonly=0) Nov 27 05:05:55 localhost ovn_controller[156436]: 2025-11-27T10:05:55Z|00177|binding|INFO|Setting lport beb4101c-445d-4f4e-8517-c39429819f92 down in Southbound Nov 27 05:05:55 localhost nova_compute[284026]: 2025-11-27 10:05:55.474 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:55 localhost nova_compute[284026]: 2025-11-27 10:05:55.476 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] 0-ms timeout __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:248#033[00m Nov 27 05:05:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:55.635 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '4', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=beb4101c-445d-4f4e-8517-c39429819f92) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:55.636 162092 INFO neutron.agent.ovn.metadata.agent [-] Port beb4101c-445d-4f4e-8517-c39429819f92 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:05:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:55.638 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:55.638 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ee5f9386-4ced-46a7-ae21-29793cea50b8]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:55 localhost openstack_network_exporter[244641]: ERROR 10:05:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:05:55 localhost openstack_network_exporter[244641]: ERROR 10:05:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:05:55 localhost openstack_network_exporter[244641]: ERROR 10:05:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:05:55 localhost openstack_network_exporter[244641]: Nov 27 05:05:55 localhost openstack_network_exporter[244641]: ERROR 10:05:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:05:55 localhost openstack_network_exporter[244641]: ERROR 10:05:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:05:55 localhost openstack_network_exporter[244641]: Nov 27 05:05:55 localhost nova_compute[284026]: 2025-11-27 10:05:55.832 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:56 localhost systemd[1]: var-lib-containers-storage-overlay-d9dea73b2ce7fac2e549d438af1b0e042b18f9d2f39c1f5ac92e957cd9b18eb5-merged.mount: Deactivated successfully. Nov 27 05:05:56 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-5d7e27c2548e63900a9fab5afddd6c30ecc67ba000f0f4a5a5d249db7933e736-userdata-shm.mount: Deactivated successfully. Nov 27 05:05:56 localhost ovn_controller[156436]: 2025-11-27T10:05:56Z|00178|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:05:56 localhost nova_compute[284026]: 2025-11-27 10:05:56.467 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e128 e128: 6 total, 6 up, 6 in Nov 27 05:05:56 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:05:56 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:56.873 265123 INFO neutron.agent.dhcp.agent [None req-75e33493-ecbd-4bc5-b062-97affcefd74c - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:05:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v229: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 40 KiB/s rd, 2.4 KiB/s wr, 53 op/s Nov 27 05:05:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e128 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:05:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:05:58 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:05:58 localhost podman[316644]: 2025-11-27 10:05:58.728441909 +0000 UTC m=+0.082959208 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:05:58 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:58.736 2 INFO neutron.agent.securitygroups_rpc [None req-cf5651f4-9b71-4184-9386-7ca143eee50e d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:05:58 localhost podman[316644]: 2025-11-27 10:05:58.761846834 +0000 UTC m=+0.116364103 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:05:58 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:05:58 localhost podman[316645]: 2025-11-27 10:05:58.833894726 +0000 UTC m=+0.188408295 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.vendor=CentOS, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:05:58 localhost podman[316645]: 2025-11-27 10:05:58.842429427 +0000 UTC m=+0.196943026 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:05:58 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:05:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v230: 177 pgs: 177 active+clean; 145 MiB data, 769 MiB used, 41 GiB / 42 GiB avail; 35 KiB/s rd, 1.6 KiB/s wr, 45 op/s Nov 27 05:05:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:05:59 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:05:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:05:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:05:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:05:59 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev f59bc2ed-fc05-4498-a3ad-6e12e87067d0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:05:59 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev f59bc2ed-fc05-4498-a3ad-6e12e87067d0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:05:59 localhost ceph-mgr[290377]: [progress INFO root] Completed event f59bc2ed-fc05-4498-a3ad-6e12e87067d0 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:05:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:05:59 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:05:59 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:05:59.555 265123 INFO neutron.agent.linux.ip_lib [None req-a33f3a6f-6959-41d3-b63e-91fb7d068b77 - - - - - -] Device tapb2a535d9-8f cannot be used as it has no MAC address#033[00m Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.615 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost kernel: device tapb2a535d9-8f entered promiscuous mode Nov 27 05:05:59 localhost NetworkManager[5971]: [1764237959.6210] manager: (tapb2a535d9-8f): new Generic device (/org/freedesktop/NetworkManager/Devices/34) Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.621 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost ovn_controller[156436]: 2025-11-27T10:05:59Z|00179|binding|INFO|Claiming lport b2a535d9-8f34-446f-9e98-cd7e5bf99305 for this chassis. Nov 27 05:05:59 localhost ovn_controller[156436]: 2025-11-27T10:05:59Z|00180|binding|INFO|b2a535d9-8f34-446f-9e98-cd7e5bf99305: Claiming unknown Nov 27 05:05:59 localhost systemd-udevd[316737]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:59.654 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=b2a535d9-8f34-446f-9e98-cd7e5bf99305) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:05:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:59.656 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b2a535d9-8f34-446f-9e98-cd7e5bf99305 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:59.658 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:05:59 localhost ovn_controller[156436]: 2025-11-27T10:05:59Z|00181|binding|INFO|Setting lport b2a535d9-8f34-446f-9e98-cd7e5bf99305 ovn-installed in OVS Nov 27 05:05:59 localhost ovn_controller[156436]: 2025-11-27T10:05:59Z|00182|binding|INFO|Setting lport b2a535d9-8f34-446f-9e98-cd7e5bf99305 up in Southbound Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:05:59.661 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a5cc66a1-4f2f-45f8-97e4-5c7314ad739e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.662 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:05:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost journal[232028]: ethtool ioctl error on tapb2a535d9-8f: No such device Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.695 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.723 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost nova_compute[284026]: 2025-11-27 10:05:59.915 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:05:59 localhost neutron_sriov_agent[258162]: 2025-11-27 10:05:59.943 2 INFO neutron.agent.securitygroups_rpc [None req-32fd7465-d5d0-42e2-89d0-92129ebfaf8b b67e89fdaf044cd7a572546dcf50de71 b090b766aebb4ed89dfb3158699f8c2e - - default default] Security group member updated ['e279e252-5eb4-4077-9fe3-4b12dcfa3db1']#033[00m Nov 27 05:06:00 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:06:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:06:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:06:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:00.409 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:59Z, description=, device_id=bd47f427-be56-4aa6-9bd1-aae0111bbb77, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=729834fb-ca25-438c-82a5-7a6aa7d4a1fc, ip_allocation=immediate, mac_address=fa:16:3e:e1:62:8c, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1293, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:00Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:00 localhost nova_compute[284026]: 2025-11-27 10:06:00.477 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:00 localhost podman[316826]: Nov 27 05:06:00 localhost podman[316826]: 2025-11-27 10:06:00.495071482 +0000 UTC m=+0.095323583 container create 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:06:00 localhost systemd[1]: Started libpod-conmon-90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84.scope. Nov 27 05:06:00 localhost podman[316826]: 2025-11-27 10:06:00.446859747 +0000 UTC m=+0.047111888 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:00 localhost systemd[1]: Started libcrun container. Nov 27 05:06:00 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2ab127c3d10ce5b09979795985f93bb5e522adf34aff5b57f13bc4ed9a470e2f/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:00 localhost podman[316826]: 2025-11-27 10:06:00.577553226 +0000 UTC m=+0.177805327 container init 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:00 localhost podman[316826]: 2025-11-27 10:06:00.588252296 +0000 UTC m=+0.188504397 container start 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:00 localhost dnsmasq[316844]: started, version 2.85 cachesize 150 Nov 27 05:06:00 localhost dnsmasq[316844]: DNS service limited to local subnets Nov 27 05:06:00 localhost dnsmasq[316844]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:00 localhost dnsmasq[316844]: warning: no upstream servers configured Nov 27 05:06:00 localhost dnsmasq[316844]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:00.645 265123 INFO neutron.agent.dhcp.agent [None req-a33f3a6f-6959-41d3-b63e-91fb7d068b77 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:59Z, description=, device_id=f2bdfdc8-cd21-4a09-b0d2-551786f7e7ea, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3cdd3915-b722-4c5d-a203-475f97ad7dc2, ip_allocation=immediate, mac_address=fa:16:3e:75:b6:a1, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=20, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['abac8a3b-306b-4175-8140-595c5470ead9'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:57Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=False, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1291, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:59Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:00 localhost nova_compute[284026]: 2025-11-27 10:06:00.836 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:00 localhost dnsmasq[316844]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:06:00 localhost podman[316877]: 2025-11-27 10:06:00.881406797 +0000 UTC m=+0.102929689 container kill 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:00 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:00 localhost podman[316889]: 2025-11-27 10:06:00.897547405 +0000 UTC m=+0.059223476 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:01.053 265123 INFO neutron.agent.dhcp.agent [None req-d42e1637-1c07-4327-a997-a9c176eacd3c - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v231: 177 pgs: 177 active+clean; 145 MiB data, 773 MiB used, 41 GiB / 42 GiB avail; 36 KiB/s rd, 1.5 KiB/s wr, 47 op/s Nov 27 05:06:01 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:01.574 2 INFO neutron.agent.securitygroups_rpc [None req-71eb9e77-b580-471c-aab4-afccbe11d038 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:01.582 265123 INFO neutron.agent.dhcp.agent [None req-9f5ac448-e8a5-49dd-b74d-df99f73c3842 - - - - - -] DHCP configuration for ports {'3cdd3915-b722-4c5d-a203-475f97ad7dc2', '729834fb-ca25-438c-82a5-7a6aa7d4a1fc'} is completed#033[00m Nov 27 05:06:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:01.803 265123 INFO neutron.agent.linux.ip_lib [None req-7d41095f-fcb3-4ba2-9598-b3b7aec29fa6 - - - - - -] Device tap4d9c5929-19 cannot be used as it has no MAC address#033[00m Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.825 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost kernel: device tap4d9c5929-19 entered promiscuous mode Nov 27 05:06:01 localhost NetworkManager[5971]: [1764237961.8341] manager: (tap4d9c5929-19): new Generic device (/org/freedesktop/NetworkManager/Devices/35) Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.835 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost ovn_controller[156436]: 2025-11-27T10:06:01Z|00183|binding|INFO|Claiming lport 4d9c5929-1943-421d-b3c6-6e11fbd3203e for this chassis. Nov 27 05:06:01 localhost ovn_controller[156436]: 2025-11-27T10:06:01Z|00184|binding|INFO|4d9c5929-1943-421d-b3c6-6e11fbd3203e: Claiming unknown Nov 27 05:06:01 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:01.853 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-43ed24e8-0ac2-4381-bb57-f89abc147d21', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-43ed24e8-0ac2-4381-bb57-f89abc147d21', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f4460e56db744027b7a29a83ceb92d47', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=aa391bf4-f7bd-4bdf-96c9-4158f06f69bb, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4d9c5929-1943-421d-b3c6-6e11fbd3203e) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:01 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:01.856 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4d9c5929-1943-421d-b3c6-6e11fbd3203e in datapath 43ed24e8-0ac2-4381-bb57-f89abc147d21 bound to our chassis#033[00m Nov 27 05:06:01 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:01.858 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 43ed24e8-0ac2-4381-bb57-f89abc147d21 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:01.859 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c60aed05-6397-4049-970c-9a9a1774c874]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.869 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost ovn_controller[156436]: 2025-11-27T10:06:01Z|00185|binding|INFO|Setting lport 4d9c5929-1943-421d-b3c6-6e11fbd3203e ovn-installed in OVS Nov 27 05:06:01 localhost ovn_controller[156436]: 2025-11-27T10:06:01Z|00186|binding|INFO|Setting lport 4d9c5929-1943-421d-b3c6-6e11fbd3203e up in Southbound Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.873 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.876 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost journal[232028]: ethtool ioctl error on tap4d9c5929-19: No such device Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.906 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost nova_compute[284026]: 2025-11-27 10:06:01.934 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:01 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:01.986 2 INFO neutron.agent.securitygroups_rpc [None req-e937c52f-8744-4e1f-80e1-eb6bd3063814 b67e89fdaf044cd7a572546dcf50de71 b090b766aebb4ed89dfb3158699f8c2e - - default default] Security group member updated ['e279e252-5eb4-4077-9fe3-4b12dcfa3db1']#033[00m Nov 27 05:06:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e128 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:02 localhost nova_compute[284026]: 2025-11-27 10:06:02.707 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:02 localhost podman[316998]: Nov 27 05:06:02 localhost podman[316998]: 2025-11-27 10:06:02.834938053 +0000 UTC m=+0.085470446 container create 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:02 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:02.837 2 INFO neutron.agent.securitygroups_rpc [None req-c813bd49-aa30-4cef-9669-64a2f26cc0b3 88f0516a86664a828ee0e3a2264f4672 f0753b0eb35c49d1852f196ec33a2877 - - default default] Security group rule updated ['677d6692-9431-43e9-bf62-1d9fe6051ff5']#033[00m Nov 27 05:06:02 localhost systemd[1]: Started libpod-conmon-7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8.scope. Nov 27 05:06:02 localhost systemd[1]: Started libcrun container. Nov 27 05:06:02 localhost podman[316998]: 2025-11-27 10:06:02.794369005 +0000 UTC m=+0.044901448 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:02 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/00b6a5b35e85a1abd360ca5895a1c349502853c2ebd60b264e196416bc0f4e3c/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:02 localhost podman[316998]: 2025-11-27 10:06:02.906007079 +0000 UTC m=+0.156539472 container init 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:02 localhost podman[316998]: 2025-11-27 10:06:02.917097388 +0000 UTC m=+0.167629821 container start 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:02 localhost dnsmasq[317016]: started, version 2.85 cachesize 150 Nov 27 05:06:02 localhost dnsmasq[317016]: DNS service limited to local subnets Nov 27 05:06:02 localhost dnsmasq[317016]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:02 localhost dnsmasq[317016]: warning: no upstream servers configured Nov 27 05:06:02 localhost dnsmasq-dhcp[317016]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:02 localhost dnsmasq[317016]: read /var/lib/neutron/dhcp/43ed24e8-0ac2-4381-bb57-f89abc147d21/addn_hosts - 0 addresses Nov 27 05:06:02 localhost dnsmasq-dhcp[317016]: read /var/lib/neutron/dhcp/43ed24e8-0ac2-4381-bb57-f89abc147d21/host Nov 27 05:06:02 localhost dnsmasq-dhcp[317016]: read /var/lib/neutron/dhcp/43ed24e8-0ac2-4381-bb57-f89abc147d21/opts Nov 27 05:06:03 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:03.061 265123 INFO neutron.agent.dhcp.agent [None req-960b9b31-1596-4c6d-8fcd-c392ce42e3de - - - - - -] DHCP configuration for ports {'bc52358d-c278-4903-a881-df267fc79636'} is completed#033[00m Nov 27 05:06:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v232: 177 pgs: 177 active+clean; 145 MiB data, 773 MiB used, 41 GiB / 42 GiB avail; 36 KiB/s rd, 1.5 KiB/s wr, 47 op/s Nov 27 05:06:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:03.773 2 INFO neutron.agent.securitygroups_rpc [None req-0eea2017-2bc5-44db-9717-9f2414586b0f 88f0516a86664a828ee0e3a2264f4672 f0753b0eb35c49d1852f196ec33a2877 - - default default] Security group rule updated ['677d6692-9431-43e9-bf62-1d9fe6051ff5']#033[00m Nov 27 05:06:04 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:04.507 2 INFO neutron.agent.securitygroups_rpc [None req-45c075f9-bb26-4bc7-b2a0-1d071cca6a0b d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:04 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:04.509 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:05:59Z, description=, device_id=f2bdfdc8-cd21-4a09-b0d2-551786f7e7ea, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3cdd3915-b722-4c5d-a203-475f97ad7dc2, ip_allocation=immediate, mac_address=fa:16:3e:75:b6:a1, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=20, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['abac8a3b-306b-4175-8140-595c5470ead9'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:57Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=False, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1291, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:05:59Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:04 localhost systemd[1]: tmp-crun.R3OYi1.mount: Deactivated successfully. Nov 27 05:06:04 localhost podman[317035]: 2025-11-27 10:06:04.6980852 +0000 UTC m=+0.058187657 container kill 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:06:04 localhost dnsmasq[316844]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:06:04 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:04.940 265123 INFO neutron.agent.dhcp.agent [None req-2ede32c1-924d-43d4-ac86-665d92013797 - - - - - -] DHCP configuration for ports {'3cdd3915-b722-4c5d-a203-475f97ad7dc2'} is completed#033[00m Nov 27 05:06:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v233: 177 pgs: 177 active+clean; 145 MiB data, 773 MiB used, 41 GiB / 42 GiB avail; 36 KiB/s rd, 1.5 KiB/s wr, 47 op/s Nov 27 05:06:05 localhost nova_compute[284026]: 2025-11-27 10:06:05.409 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:05 localhost nova_compute[284026]: 2025-11-27 10:06:05.478 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:05 localhost nova_compute[284026]: 2025-11-27 10:06:05.875 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:06 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:06.551 2 INFO neutron.agent.securitygroups_rpc [None req-0e9dcc18-724c-4df5-9012-bc791538e9c6 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:06 localhost podman[317074]: 2025-11-27 10:06:06.823999386 +0000 UTC m=+0.077771168 container kill 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:06 localhost dnsmasq[317016]: exiting on receipt of SIGTERM Nov 27 05:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:06:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:06:06 localhost systemd[1]: tmp-crun.s6rPy2.mount: Deactivated successfully. Nov 27 05:06:06 localhost systemd[1]: libpod-7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8.scope: Deactivated successfully. Nov 27 05:06:06 localhost podman[317100]: 2025-11-27 10:06:06.908921126 +0000 UTC m=+0.050601241 container died 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:06 localhost systemd[1]: tmp-crun.PMZT2l.mount: Deactivated successfully. Nov 27 05:06:06 localhost podman[317098]: 2025-11-27 10:06:06.952847536 +0000 UTC m=+0.100659727 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, managed_by=edpm_ansible, config_id=edpm, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 05:06:06 localhost podman[317098]: 2025-11-27 10:06:06.963902396 +0000 UTC m=+0.111714557 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:06:06 localhost podman[317101]: 2025-11-27 10:06:06.991332059 +0000 UTC m=+0.139761437 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:06:06 localhost dnsmasq[316844]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:06 localhost podman[317133]: 2025-11-27 10:06:06.994255138 +0000 UTC m=+0.111443100 container kill 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:06:07 localhost podman[317104]: 2025-11-27 10:06:07.043580625 +0000 UTC m=+0.181514278 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, name=ubi9-minimal, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., architecture=x86_64, com.redhat.component=ubi9-minimal-container, release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, distribution-scope=public, vcs-type=git, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., managed_by=edpm_ansible, config_id=edpm, io.openshift.expose-services=, version=9.6, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7) Nov 27 05:06:07 localhost podman[317104]: 2025-11-27 10:06:07.054730096 +0000 UTC m=+0.192663759 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., container_name=openstack_network_exporter, io.buildah.version=1.33.7, release=1755695350, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41, version=9.6, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, maintainer=Red Hat, Inc., vcs-type=git) Nov 27 05:06:07 localhost podman[317101]: 2025-11-27 10:06:07.063386381 +0000 UTC m=+0.211815809 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:06:07 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:06:07 localhost podman[317100]: 2025-11-27 10:06:07.115747149 +0000 UTC m=+0.257427224 container remove 7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-43ed24e8-0ac2-4381-bb57-f89abc147d21, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0) Nov 27 05:06:07 localhost systemd[1]: libpod-conmon-7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8.scope: Deactivated successfully. Nov 27 05:06:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v234: 177 pgs: 177 active+clean; 145 MiB data, 773 MiB used, 41 GiB / 42 GiB avail; 6.8 KiB/s rd, 195 B/s wr, 8 op/s Nov 27 05:06:07 localhost kernel: device tap4d9c5929-19 left promiscuous mode Nov 27 05:06:07 localhost nova_compute[284026]: 2025-11-27 10:06:07.159 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:07 localhost ovn_controller[156436]: 2025-11-27T10:06:07Z|00187|binding|INFO|Releasing lport 4d9c5929-1943-421d-b3c6-6e11fbd3203e from this chassis (sb_readonly=0) Nov 27 05:06:07 localhost ovn_controller[156436]: 2025-11-27T10:06:07Z|00188|binding|INFO|Setting lport 4d9c5929-1943-421d-b3c6-6e11fbd3203e down in Southbound Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.170 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-43ed24e8-0ac2-4381-bb57-f89abc147d21', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-43ed24e8-0ac2-4381-bb57-f89abc147d21', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f4460e56db744027b7a29a83ceb92d47', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=aa391bf4-f7bd-4bdf-96c9-4158f06f69bb, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4d9c5929-1943-421d-b3c6-6e11fbd3203e) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.171 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4d9c5929-1943-421d-b3c6-6e11fbd3203e in datapath 43ed24e8-0ac2-4381-bb57-f89abc147d21 unbound from our chassis#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.172 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 43ed24e8-0ac2-4381-bb57-f89abc147d21, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.173 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4f174f0c-6a59-476e-8fec-c4f9bea28d13]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:07 localhost nova_compute[284026]: 2025-11-27 10:06:07.180 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:07 localhost nova_compute[284026]: 2025-11-27 10:06:07.184 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:07 localhost ovn_controller[156436]: 2025-11-27T10:06:07Z|00189|binding|INFO|Releasing lport b2a535d9-8f34-446f-9e98-cd7e5bf99305 from this chassis (sb_readonly=0) Nov 27 05:06:07 localhost kernel: device tapb2a535d9-8f left promiscuous mode Nov 27 05:06:07 localhost ovn_controller[156436]: 2025-11-27T10:06:07Z|00190|binding|INFO|Setting lport b2a535d9-8f34-446f-9e98-cd7e5bf99305 down in Southbound Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.192 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=b2a535d9-8f34-446f-9e98-cd7e5bf99305) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.193 162092 INFO neutron.agent.ovn.metadata.agent [-] Port b2a535d9-8f34-446f-9e98-cd7e5bf99305 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.194 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:06:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:07.195 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[7dd7bc24-e1a9-4f35-a042-f202fe44a8c2]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:07 localhost nova_compute[284026]: 2025-11-27 10:06:07.202 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e128 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:07 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:07.622 265123 INFO neutron.agent.dhcp.agent [None req-8d5ad9dc-57ea-491e-a5bf-919732bbbd51 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:06:07 localhost dnsmasq[316844]: exiting on receipt of SIGTERM Nov 27 05:06:07 localhost podman[317227]: 2025-11-27 10:06:07.742659131 +0000 UTC m=+0.058433154 container kill 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:07 localhost systemd[1]: libpod-90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84.scope: Deactivated successfully. Nov 27 05:06:07 localhost podman[317243]: 2025-11-27 10:06:07.803638933 +0000 UTC m=+0.042484262 container died 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:06:07 localhost systemd[1]: var-lib-containers-storage-overlay-00b6a5b35e85a1abd360ca5895a1c349502853c2ebd60b264e196416bc0f4e3c-merged.mount: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-7c4f497c1f1aacc1ca662becfe7894ddf4fc30b2dec2ee52cffd975b61f457c8-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: run-netns-qdhcp\x2d43ed24e8\x2d0ac2\x2d4381\x2dbb57\x2df89abc147d21.mount: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: tmp-crun.A1dOqO.mount: Deactivated successfully. Nov 27 05:06:07 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:07 localhost podman[317243]: 2025-11-27 10:06:07.862904758 +0000 UTC m=+0.101750097 container remove 90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:06:07 localhost ovn_controller[156436]: 2025-11-27T10:06:07Z|00191|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:06:07 localhost systemd[1]: libpod-conmon-90d7ea192aaf6b96c748237d0fb7cb1caf01968cab9a9971377de54fd3f26a84.scope: Deactivated successfully. Nov 27 05:06:07 localhost nova_compute[284026]: 2025-11-27 10:06:07.944 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:08.143 265123 INFO neutron.agent.dhcp.agent [None req-e64d25ea-1627-4562-b4ea-bcd7660e3beb - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:06:08 localhost podman[242678]: time="2025-11-27T10:06:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:06:08 localhost podman[242678]: @ - - [27/Nov/2025:10:06:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:06:08 localhost systemd[1]: var-lib-containers-storage-overlay-2ab127c3d10ce5b09979795985f93bb5e522adf34aff5b57f13bc4ed9a470e2f-merged.mount: Deactivated successfully. Nov 27 05:06:08 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:06:08 localhost podman[242678]: @ - - [27/Nov/2025:10:06:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19736 "" "Go-http-client/1.1" Nov 27 05:06:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v235: 177 pgs: 177 active+clean; 145 MiB data, 773 MiB used, 41 GiB / 42 GiB avail; 5.9 KiB/s rd, 170 B/s wr, 7 op/s Nov 27 05:06:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:09.616 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:06:09 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:09.779 2 INFO neutron.agent.securitygroups_rpc [req-d84d9d27-067a-4b82-89f3-902bb135fcdc req-838e5a2f-c00a-4039-8868-ea1f22095ac7 88f0516a86664a828ee0e3a2264f4672 f0753b0eb35c49d1852f196ec33a2877 - - default default] Security group member updated ['677d6692-9431-43e9-bf62-1d9fe6051ff5']#033[00m Nov 27 05:06:10 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:10.306 265123 INFO neutron.agent.linux.ip_lib [None req-d9278723-e1f1-43e8-aaed-d0099628a1f2 - - - - - -] Device tapc4f8f726-1e cannot be used as it has no MAC address#033[00m Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.328 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost kernel: device tapc4f8f726-1e entered promiscuous mode Nov 27 05:06:10 localhost NetworkManager[5971]: [1764237970.3384] manager: (tapc4f8f726-1e): new Generic device (/org/freedesktop/NetworkManager/Devices/36) Nov 27 05:06:10 localhost ovn_controller[156436]: 2025-11-27T10:06:10Z|00192|binding|INFO|Claiming lport c4f8f726-1e25-4421-bf8b-ec517e35bed8 for this chassis. Nov 27 05:06:10 localhost ovn_controller[156436]: 2025-11-27T10:06:10Z|00193|binding|INFO|c4f8f726-1e25-4421-bf8b-ec517e35bed8: Claiming unknown Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.339 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost systemd-udevd[317280]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:06:10 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:10.349 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe84:f379/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=c4f8f726-1e25-4421-bf8b-ec517e35bed8) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:10 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:10.351 162092 INFO neutron.agent.ovn.metadata.agent [-] Port c4f8f726-1e25-4421-bf8b-ec517e35bed8 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:06:10 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:10.355 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 43d2773d-aab2-4a14-ad64-45f43f7e4675 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:10 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:10.355 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:10 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:10.356 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4fd29cbf-5c1b-416e-9eb6-fd90f4419a70]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost ovn_controller[156436]: 2025-11-27T10:06:10Z|00194|binding|INFO|Setting lport c4f8f726-1e25-4421-bf8b-ec517e35bed8 ovn-installed in OVS Nov 27 05:06:10 localhost ovn_controller[156436]: 2025-11-27T10:06:10Z|00195|binding|INFO|Setting lport c4f8f726-1e25-4421-bf8b-ec517e35bed8 up in Southbound Nov 27 05:06:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:10.380 2 INFO neutron.agent.securitygroups_rpc [None req-2f9f276f-41d3-41d9-a2c7-cdbbba3ed55a d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.381 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost journal[232028]: ethtool ioctl error on tapc4f8f726-1e: No such device Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.422 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.451 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.479 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:10 localhost nova_compute[284026]: 2025-11-27 10:06:10.905 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v236: 177 pgs: 177 active+clean; 192 MiB data, 837 MiB used, 41 GiB / 42 GiB avail; 23 KiB/s rd, 1.8 MiB/s wr, 35 op/s Nov 27 05:06:11 localhost podman[317351]: Nov 27 05:06:11 localhost podman[317351]: 2025-11-27 10:06:11.238075564 +0000 UTC m=+0.087755889 container create 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e129 e129: 6 total, 6 up, 6 in Nov 27 05:06:11 localhost systemd[1]: Started libpod-conmon-976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4.scope. Nov 27 05:06:11 localhost systemd[1]: tmp-crun.fmGuNP.mount: Deactivated successfully. Nov 27 05:06:11 localhost systemd[1]: Started libcrun container. Nov 27 05:06:11 localhost podman[317351]: 2025-11-27 10:06:11.19585696 +0000 UTC m=+0.045537355 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:11 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/9437093478dbd7afcc54b1553f244bcd1523c70b43b140c5e930fa4ccb49d48b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:11 localhost podman[317351]: 2025-11-27 10:06:11.304584325 +0000 UTC m=+0.154264650 container init 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:06:11 localhost podman[317351]: 2025-11-27 10:06:11.312893501 +0000 UTC m=+0.162573816 container start 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:06:11 localhost dnsmasq[317369]: started, version 2.85 cachesize 150 Nov 27 05:06:11 localhost dnsmasq[317369]: DNS service limited to local subnets Nov 27 05:06:11 localhost dnsmasq[317369]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:11 localhost dnsmasq[317369]: warning: no upstream servers configured Nov 27 05:06:11 localhost dnsmasq[317369]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:11.322 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '2', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:11.324 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:11.327 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 43d2773d-aab2-4a14-ad64-45f43f7e4675 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:11.327 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:11.328 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[12021e81-646b-4a7c-b383-5bfcc02b25bd]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:11 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:11.480 265123 INFO neutron.agent.dhcp.agent [None req-edcf2abc-beef-4542-b328-cb843359f8c0 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:11 localhost systemd[1]: virtsecretd.service: Deactivated successfully. Nov 27 05:06:11 localhost dnsmasq[317369]: exiting on receipt of SIGTERM Nov 27 05:06:11 localhost podman[317388]: 2025-11-27 10:06:11.680017675 +0000 UTC m=+0.061515267 container kill 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:11 localhost systemd[1]: libpod-976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4.scope: Deactivated successfully. Nov 27 05:06:11 localhost podman[317402]: 2025-11-27 10:06:11.743802272 +0000 UTC m=+0.048193346 container died 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:06:11 localhost podman[317402]: 2025-11-27 10:06:11.773873697 +0000 UTC m=+0.078264731 container cleanup 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:11 localhost systemd[1]: libpod-conmon-976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4.scope: Deactivated successfully. Nov 27 05:06:11 localhost podman[317404]: 2025-11-27 10:06:11.796910691 +0000 UTC m=+0.094429459 container remove 976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:06:12 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:12.219 2 INFO neutron.agent.securitygroups_rpc [None req-8e787331-71a3-41d2-b063-6f293287ab44 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:12 localhost systemd[1]: var-lib-containers-storage-overlay-9437093478dbd7afcc54b1553f244bcd1523c70b43b140c5e930fa4ccb49d48b-merged.mount: Deactivated successfully. Nov 27 05:06:12 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-976abe0286c237747e01bdc9eb70487497866eee8392f7df378e9676d1df14d4-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e129 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:13 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:13.119 2 INFO neutron.agent.securitygroups_rpc [None req-dfad064d-f83b-47e8-a12f-76103410de79 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v238: 177 pgs: 177 active+clean; 192 MiB data, 837 MiB used, 41 GiB / 42 GiB avail; 21 KiB/s rd, 2.1 MiB/s wr, 33 op/s Nov 27 05:06:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e130 e130: 6 total, 6 up, 6 in Nov 27 05:06:13 localhost nova_compute[284026]: 2025-11-27 10:06:13.372 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:13 localhost podman[317484]: Nov 27 05:06:13 localhost podman[317484]: 2025-11-27 10:06:13.393838877 +0000 UTC m=+0.093743460 container create fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:06:13 localhost systemd[1]: Started libpod-conmon-fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3.scope. Nov 27 05:06:13 localhost systemd[1]: Started libcrun container. Nov 27 05:06:13 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d0600399f5eab15a51a0b9242ff9ac02d0a89df2d30282a5b382b76cb919d377/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:13 localhost podman[317484]: 2025-11-27 10:06:13.350826872 +0000 UTC m=+0.050731475 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:13 localhost podman[317484]: 2025-11-27 10:06:13.457543573 +0000 UTC m=+0.157448146 container init fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:06:13 localhost podman[317484]: 2025-11-27 10:06:13.469690521 +0000 UTC m=+0.169595104 container start fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:06:13 localhost dnsmasq[317502]: started, version 2.85 cachesize 150 Nov 27 05:06:13 localhost dnsmasq[317502]: DNS service limited to local subnets Nov 27 05:06:13 localhost dnsmasq[317502]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:13 localhost dnsmasq[317502]: warning: no upstream servers configured Nov 27 05:06:13 localhost dnsmasq-dhcp[317502]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:13 localhost dnsmasq[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:13 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:13 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:13.529 265123 INFO neutron.agent.dhcp.agent [None req-e54f0c65-78fe-4e46-8aab-223f105f3551 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:12Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=2d22427d-8afc-483d-9219-3856deaadee3, ip_allocation=immediate, mac_address=fa:16:3e:42:af:74, name=tempest-NetworksTestDHCPv6-362580259, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=23, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['de775952-1928-4f51-b130-3205b6c83187', 'f9ab8c2f-039f-407a-807c-640decefaf84'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:10Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1323, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:12Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:13.656 265123 INFO neutron.agent.dhcp.agent [None req-97f1bccf-df22-46b1-9dc3-f643fb414eff - - - - - -] DHCP configuration for ports {'c4f8f726-1e25-4421-bf8b-ec517e35bed8', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:13 localhost dnsmasq[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:06:13 localhost podman[317520]: 2025-11-27 10:06:13.786182374 +0000 UTC m=+0.065850404 container kill fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:06:13 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:13 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:14.253 265123 INFO neutron.agent.dhcp.agent [None req-f480e8b8-9aac-4d94-a0bc-8680aa8f8523 - - - - - -] DHCP configuration for ports {'2d22427d-8afc-483d-9219-3856deaadee3'} is completed#033[00m Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:14.860 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:14Z, description=, device_id=0697acd0-6e7d-466f-aa42-cdd764fd029c, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=86bfbd8d-1bd7-450d-b64f-8a3f7ed181e0, ip_allocation=immediate, mac_address=fa:16:3e:3d:ae:28, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1327, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:14Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:15.081 2 INFO neutron.agent.securitygroups_rpc [None req-801ee41f-78bf-4a0d-b43c-54585803b27c 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:15 localhost systemd[1]: tmp-crun.QT6Y34.mount: Deactivated successfully. Nov 27 05:06:15 localhost podman[317558]: 2025-11-27 10:06:15.108284697 +0000 UTC m=+0.056891182 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:06:15 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:06:15 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:15 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:06:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:06:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v240: 177 pgs: 177 active+clean; 192 MiB data, 837 MiB used, 41 GiB / 42 GiB avail; 26 KiB/s rd, 2.7 MiB/s wr, 42 op/s Nov 27 05:06:15 localhost systemd[1]: tmp-crun.0m56EO.mount: Deactivated successfully. Nov 27 05:06:15 localhost podman[317574]: 2025-11-27 10:06:15.20845079 +0000 UTC m=+0.079466523 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:06:15 localhost podman[317574]: 2025-11-27 10:06:15.243776027 +0000 UTC m=+0.114791690 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:06:15 localhost podman[317572]: 2025-11-27 10:06:15.253789068 +0000 UTC m=+0.124823222 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 05:06:15 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:06:15 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e131 e131: 6 total, 6 up, 6 in Nov 27 05:06:15 localhost podman[317572]: 2025-11-27 10:06:15.314938565 +0000 UTC m=+0.185972679 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:06:15 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:06:15 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:15.436 265123 INFO neutron.agent.dhcp.agent [None req-3da93cd3-e207-4bee-b0f8-5980a22f843e - - - - - -] DHCP configuration for ports {'86bfbd8d-1bd7-450d-b64f-8a3f7ed181e0'} is completed#033[00m Nov 27 05:06:15 localhost nova_compute[284026]: 2025-11-27 10:06:15.481 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:15 localhost dnsmasq[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:15 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:15 localhost podman[317645]: 2025-11-27 10:06:15.633884734 +0000 UTC m=+0.051039743 container kill fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:06:15 localhost dnsmasq-dhcp[317502]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:15.721 2 INFO neutron.agent.securitygroups_rpc [None req-6ef33e67-1db7-41b5-bc4a-c928d4ae766e d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:15 localhost nova_compute[284026]: 2025-11-27 10:06:15.908 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e132 e132: 6 total, 6 up, 6 in Nov 27 05:06:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:16.890 2 INFO neutron.agent.securitygroups_rpc [req-006b0ae6-282f-459f-a1e1-899b371f6843 req-ced2cd52-dafa-46b9-b3a9-355994addda9 c2db1ef6590d40f88d1037c6c10ab3a6 891efae95c00451a835153cbe11a1ee8 - - default default] Security group rule updated ['61e529b0-9f0f-4389-94a1-3b20e71c9c14']#033[00m Nov 27 05:06:17 localhost dnsmasq[317502]: exiting on receipt of SIGTERM Nov 27 05:06:17 localhost podman[317682]: 2025-11-27 10:06:17.147613406 +0000 UTC m=+0.061759674 container kill fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:06:17 localhost systemd[1]: libpod-fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3.scope: Deactivated successfully. Nov 27 05:06:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v243: 177 pgs: 177 active+clean; 192 MiB data, 837 MiB used, 41 GiB / 42 GiB avail; 37 KiB/s rd, 2.4 KiB/s wr, 50 op/s Nov 27 05:06:17 localhost podman[317695]: 2025-11-27 10:06:17.216873512 +0000 UTC m=+0.055184086 container died fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:06:17 localhost podman[317695]: 2025-11-27 10:06:17.248724394 +0000 UTC m=+0.087034928 container cleanup fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:06:17 localhost systemd[1]: libpod-conmon-fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3.scope: Deactivated successfully. Nov 27 05:06:17 localhost podman[317697]: 2025-11-27 10:06:17.298668358 +0000 UTC m=+0.127062453 container remove fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:06:17 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:17.329 2 INFO neutron.agent.securitygroups_rpc [None req-6b42cfc2-5f91-4402-b9a1-896f7ddc5321 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e132 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:18 localhost podman[317773]: Nov 27 05:06:18 localhost podman[317773]: 2025-11-27 10:06:18.109540252 +0000 UTC m=+0.089241608 container create 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:06:18 localhost systemd[1]: Started libpod-conmon-2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e.scope. Nov 27 05:06:18 localhost systemd[1]: var-lib-containers-storage-overlay-d0600399f5eab15a51a0b9242ff9ac02d0a89df2d30282a5b382b76cb919d377-merged.mount: Deactivated successfully. Nov 27 05:06:18 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-fed028bccc0d92d7eaf6a6cc34ab3b3b598a755452597209f7e356a75ecb32d3-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:18 localhost systemd[1]: Started libcrun container. Nov 27 05:06:18 localhost podman[317773]: 2025-11-27 10:06:18.069194859 +0000 UTC m=+0.048896245 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:18 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/84d357df7d25ec05652c1f898ff2d924bfc431b515c2d7c5b93d88b01e45ecae/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:18 localhost podman[317773]: 2025-11-27 10:06:18.180910455 +0000 UTC m=+0.160611811 container init 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:06:18 localhost podman[317773]: 2025-11-27 10:06:18.190104064 +0000 UTC m=+0.169805450 container start 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:06:18 localhost dnsmasq[317791]: started, version 2.85 cachesize 150 Nov 27 05:06:18 localhost dnsmasq[317791]: DNS service limited to local subnets Nov 27 05:06:18 localhost dnsmasq[317791]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:18 localhost dnsmasq[317791]: warning: no upstream servers configured Nov 27 05:06:18 localhost dnsmasq[317791]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e133 e133: 6 total, 6 up, 6 in Nov 27 05:06:18 localhost dnsmasq[317791]: exiting on receipt of SIGTERM Nov 27 05:06:18 localhost podman[317810]: 2025-11-27 10:06:18.534252166 +0000 UTC m=+0.051099685 container kill 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:06:18 localhost systemd[1]: libpod-2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e.scope: Deactivated successfully. Nov 27 05:06:18 localhost podman[317826]: 2025-11-27 10:06:18.6085787 +0000 UTC m=+0.056967295 container died 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:18 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:18.619 2 INFO neutron.agent.securitygroups_rpc [req-b433ed22-88cd-4f1e-8bf1-c7ce2f96021a req-f5e174bb-3e88-4fd2-9026-d438d7e51625 c2db1ef6590d40f88d1037c6c10ab3a6 891efae95c00451a835153cbe11a1ee8 - - default default] Security group rule updated ['61e529b0-9f0f-4389-94a1-3b20e71c9c14']#033[00m Nov 27 05:06:18 localhost podman[317826]: 2025-11-27 10:06:18.633961217 +0000 UTC m=+0.082349802 container cleanup 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:06:18 localhost systemd[1]: libpod-conmon-2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e.scope: Deactivated successfully. Nov 27 05:06:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:18.661 265123 INFO neutron.agent.dhcp.agent [None req-65ff1330-45d5-4892-b10f-8f4cf7f08d57 - - - - - -] DHCP configuration for ports {'c4f8f726-1e25-4421-bf8b-ec517e35bed8', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:18 localhost podman[317825]: 2025-11-27 10:06:18.685340939 +0000 UTC m=+0.130533946 container remove 2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:18 localhost nova_compute[284026]: 2025-11-27 10:06:18.696 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:18 localhost ovn_controller[156436]: 2025-11-27T10:06:18Z|00196|binding|INFO|Releasing lport c4f8f726-1e25-4421-bf8b-ec517e35bed8 from this chassis (sb_readonly=0) Nov 27 05:06:18 localhost ovn_controller[156436]: 2025-11-27T10:06:18Z|00197|binding|INFO|Setting lport c4f8f726-1e25-4421-bf8b-ec517e35bed8 down in Southbound Nov 27 05:06:18 localhost kernel: device tapc4f8f726-1e left promiscuous mode Nov 27 05:06:18 localhost nova_compute[284026]: 2025-11-27 10:06:18.724 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:18.736 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe84:f379/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '4', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=c4f8f726-1e25-4421-bf8b-ec517e35bed8) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:18 localhost nova_compute[284026]: 2025-11-27 10:06:18.736 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:18.738 162092 INFO neutron.agent.ovn.metadata.agent [-] Port c4f8f726-1e25-4421-bf8b-ec517e35bed8 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:06:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:18.741 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:18 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:18.742 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[d3f7862a-c31a-4219-9af2-fed6a88470c3]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:19 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:19.047 265123 INFO neutron.agent.dhcp.agent [None req-047b347c-7c33-4091-b38d-8287b7ea3559 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:06:19 localhost systemd[1]: tmp-crun.FTmhO0.mount: Deactivated successfully. Nov 27 05:06:19 localhost systemd[1]: var-lib-containers-storage-overlay-84d357df7d25ec05652c1f898ff2d924bfc431b515c2d7c5b93d88b01e45ecae-merged.mount: Deactivated successfully. Nov 27 05:06:19 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2448dd36c0f94170f454f471d88c2eb2fb0bc02ea97602aead4af24784cee14e-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:19 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:06:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v245: 177 pgs: 177 active+clean; 192 MiB data, 837 MiB used, 41 GiB / 42 GiB avail; 37 KiB/s rd, 2.4 KiB/s wr, 50 op/s Nov 27 05:06:19 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:19.235 2 INFO neutron.agent.securitygroups_rpc [None req-191bd1e4-28a5-4e0e-9ffd-8c44772afc03 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:19 localhost podman[317869]: 2025-11-27 10:06:19.370618672 +0000 UTC m=+0.062035551 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:19 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:19 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:19 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e134 e134: 6 total, 6 up, 6 in Nov 27 05:06:20 localhost nova_compute[284026]: 2025-11-27 10:06:20.484 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:20 localhost nova_compute[284026]: 2025-11-27 10:06:20.912 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:21.135 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 2001:db8::f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '6', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:21.136 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:21.139 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:21.141 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b5591dae-c674-4189-a84c-04d7e58da4f1]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v247: 177 pgs: 177 active+clean; 192 MiB data, 838 MiB used, 41 GiB / 42 GiB avail; 2.2 MiB/s rd, 34 KiB/s wr, 187 op/s Nov 27 05:06:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e135 e135: 6 total, 6 up, 6 in Nov 27 05:06:21 localhost nova_compute[284026]: 2025-11-27 10:06:21.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:21 localhost nova_compute[284026]: 2025-11-27 10:06:21.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:21 localhost nova_compute[284026]: 2025-11-27 10:06:21.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:21 localhost nova_compute[284026]: 2025-11-27 10:06:21.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:21 localhost nova_compute[284026]: 2025-11-27 10:06:21.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.371 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=14, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=13) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.373 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 5 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:06:22 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:22.748 2 INFO neutron.agent.securitygroups_rpc [None req-64532595-2157-4253-b25e-199cc0c46db9 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['0ef4ee6e-20bb-4dac-975c-9b8981704a89']#033[00m Nov 27 05:06:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e136 e136: 6 total, 6 up, 6 in Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.783 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e136 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:22.869 265123 INFO neutron.agent.linux.ip_lib [None req-ed2e9333-6891-44b5-b31c-5531b8703d8a - - - - - -] Device tap226ebddc-31 cannot be used as it has no MAC address#033[00m Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.894 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost kernel: device tap226ebddc-31 entered promiscuous mode Nov 27 05:06:22 localhost NetworkManager[5971]: [1764237982.9023] manager: (tap226ebddc-31): new Generic device (/org/freedesktop/NetworkManager/Devices/37) Nov 27 05:06:22 localhost ovn_controller[156436]: 2025-11-27T10:06:22Z|00198|binding|INFO|Claiming lport 226ebddc-31ad-4e27-8efb-98e9f84cfce2 for this chassis. Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.902 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost ovn_controller[156436]: 2025-11-27T10:06:22Z|00199|binding|INFO|226ebddc-31ad-4e27-8efb-98e9f84cfce2: Claiming unknown Nov 27 05:06:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:22.899 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:21Z, description=, device_id=75827dae-4c3c-4846-bfce-68e9600f732d, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=77607837-9d07-486a-bd4c-ab31cebdf441, ip_allocation=immediate, mac_address=fa:16:3e:f1:0f:84, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1359, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:21Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:22 localhost systemd-udevd[317901]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:06:22 localhost ovn_controller[156436]: 2025-11-27T10:06:22Z|00200|binding|INFO|Setting lport 226ebddc-31ad-4e27-8efb-98e9f84cfce2 ovn-installed in OVS Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.915 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.920 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.947 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost ovn_controller[156436]: 2025-11-27T10:06:22Z|00201|binding|INFO|Setting lport 226ebddc-31ad-4e27-8efb-98e9f84cfce2 up in Southbound Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.981 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe98:4365/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=226ebddc-31ad-4e27-8efb-98e9f84cfce2) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.983 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 226ebddc-31ad-4e27-8efb-98e9f84cfce2 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.986 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.986 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:22 localhost nova_compute[284026]: 2025-11-27 10:06:22.986 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:22 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:22.987 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[720f626a-d918-4098-9cff-75278f7a120d]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:23 localhost nova_compute[284026]: 2025-11-27 10:06:23.014 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:23 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:23.131 2 INFO neutron.agent.securitygroups_rpc [None req-45389d93-a1b8-426f-ad51-28f9f597c131 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['0ef4ee6e-20bb-4dac-975c-9b8981704a89']#033[00m Nov 27 05:06:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v250: 177 pgs: 177 active+clean; 192 MiB data, 838 MiB used, 41 GiB / 42 GiB avail; 2.6 MiB/s rd, 38 KiB/s wr, 166 op/s Nov 27 05:06:23 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:06:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:23 localhost podman[317932]: 2025-11-27 10:06:23.232264153 +0000 UTC m=+0.065423812 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:06:23 localhost systemd[1]: tmp-crun.yEXu5X.mount: Deactivated successfully. Nov 27 05:06:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:23.384 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '7', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '6', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:23.386 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:23.389 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:23.390 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:23.391 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a478e4f6-fd2f-4d7c-be55-11e7b06f2c34]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:23.628 265123 INFO neutron.agent.dhcp.agent [None req-9a2ce8ff-d2f0-4401-8fd9-77f49dd44690 - - - - - -] DHCP configuration for ports {'77607837-9d07-486a-bd4c-ab31cebdf441'} is completed#033[00m Nov 27 05:06:23 localhost podman[317992]: Nov 27 05:06:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e137 e137: 6 total, 6 up, 6 in Nov 27 05:06:23 localhost podman[317992]: 2025-11-27 10:06:23.869438833 +0000 UTC m=+0.090239955 container create 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:23 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:23.909 2 INFO neutron.agent.securitygroups_rpc [None req-58a15fd3-60e9-45ec-ad81-dcfb1dfe6346 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:23 localhost systemd[1]: Started libpod-conmon-7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16.scope. Nov 27 05:06:23 localhost podman[317992]: 2025-11-27 10:06:23.824467735 +0000 UTC m=+0.045268857 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:23 localhost systemd[1]: Started libcrun container. Nov 27 05:06:23 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/7a055851bce9d24b386ca1a163e6af6bf44171177edb63739fcc07c060ac3faf/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:23 localhost podman[317992]: 2025-11-27 10:06:23.950196011 +0000 UTC m=+0.170997113 container init 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:06:23 localhost podman[317992]: 2025-11-27 10:06:23.96383736 +0000 UTC m=+0.184638482 container start 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:06:23 localhost dnsmasq[318010]: started, version 2.85 cachesize 150 Nov 27 05:06:23 localhost dnsmasq[318010]: DNS service limited to local subnets Nov 27 05:06:23 localhost dnsmasq[318010]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:23 localhost dnsmasq[318010]: warning: no upstream servers configured Nov 27 05:06:23 localhost dnsmasq-dhcp[318010]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:06:23 localhost dnsmasq[318010]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:23 localhost dnsmasq-dhcp[318010]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:23 localhost dnsmasq-dhcp[318010]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:24.263 265123 INFO neutron.agent.dhcp.agent [None req-bd1b5929-a9b7-4fe7-aea7-34c80caf3876 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:24 localhost dnsmasq[318010]: exiting on receipt of SIGTERM Nov 27 05:06:24 localhost podman[318028]: 2025-11-27 10:06:24.448013195 +0000 UTC m=+0.052709729 container kill 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:06:24 localhost systemd[1]: libpod-7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16.scope: Deactivated successfully. Nov 27 05:06:24 localhost podman[318040]: 2025-11-27 10:06:24.504968128 +0000 UTC m=+0.048051223 container died 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:24 localhost podman[318040]: 2025-11-27 10:06:24.552742862 +0000 UTC m=+0.095825947 container cleanup 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:06:24 localhost systemd[1]: libpod-conmon-7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16.scope: Deactivated successfully. Nov 27 05:06:24 localhost podman[318047]: 2025-11-27 10:06:24.615324957 +0000 UTC m=+0.139709365 container remove 7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:06:24 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:24.677 2 INFO neutron.agent.securitygroups_rpc [None req-6d404da1-d2a1-40be-a363-d121bddc8534 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:24 localhost nova_compute[284026]: 2025-11-27 10:06:24.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:24 localhost systemd[1]: var-lib-containers-storage-overlay-7a055851bce9d24b386ca1a163e6af6bf44171177edb63739fcc07c060ac3faf-merged.mount: Deactivated successfully. Nov 27 05:06:24 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-7b7832eaba9d92ca1014603737be26401e7793c6d57c1e7d08ff613aa9a70a16-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v252: 177 pgs: 177 active+clean; 192 MiB data, 838 MiB used, 41 GiB / 42 GiB avail; 2.7 MiB/s rd, 39 KiB/s wr, 169 op/s Nov 27 05:06:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:25.204 2 INFO neutron.agent.securitygroups_rpc [None req-3354cc38-ee69-42a1-9fd0-6b1de6389e41 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:25 localhost nova_compute[284026]: 2025-11-27 10:06:25.487 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:25.655 2 INFO neutron.agent.securitygroups_rpc [None req-5d8718b4-884b-4681-9c28-548f8dec79b5 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:25 localhost openstack_network_exporter[244641]: ERROR 10:06:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:06:25 localhost openstack_network_exporter[244641]: ERROR 10:06:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:06:25 localhost openstack_network_exporter[244641]: ERROR 10:06:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:06:25 localhost openstack_network_exporter[244641]: ERROR 10:06:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:06:25 localhost openstack_network_exporter[244641]: Nov 27 05:06:25 localhost openstack_network_exporter[244641]: ERROR 10:06:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:06:25 localhost openstack_network_exporter[244641]: Nov 27 05:06:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e138 e138: 6 total, 6 up, 6 in Nov 27 05:06:25 localhost nova_compute[284026]: 2025-11-27 10:06:25.915 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:26.058 2 INFO neutron.agent.securitygroups_rpc [None req-26db11a0-568e-408c-a3a3-0d16abdbefc1 d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:26 localhost nova_compute[284026]: 2025-11-27 10:06:26.216 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:26.518 2 INFO neutron.agent.securitygroups_rpc [None req-5e2028d9-1d05-4a19-8d89-3b5d4013407b 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:26 localhost podman[318121]: Nov 27 05:06:26 localhost podman[318121]: 2025-11-27 10:06:26.576789378 +0000 UTC m=+0.103773462 container create 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:26 localhost systemd[1]: Started libpod-conmon-7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012.scope. Nov 27 05:06:26 localhost systemd[1]: tmp-crun.z2yNcb.mount: Deactivated successfully. Nov 27 05:06:26 localhost podman[318121]: 2025-11-27 10:06:26.527311348 +0000 UTC m=+0.054295442 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:26 localhost systemd[1]: Started libcrun container. Nov 27 05:06:26 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/39b59d2ae05b8aa64d8fadb1b59468285a56578fe24c26a8833b91bc02998cf9/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:26 localhost podman[318121]: 2025-11-27 10:06:26.646684321 +0000 UTC m=+0.173668405 container init 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:26 localhost podman[318121]: 2025-11-27 10:06:26.656343673 +0000 UTC m=+0.183327757 container start 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:06:26 localhost dnsmasq[318140]: started, version 2.85 cachesize 150 Nov 27 05:06:26 localhost dnsmasq[318140]: DNS service limited to local subnets Nov 27 05:06:26 localhost dnsmasq[318140]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:26 localhost dnsmasq[318140]: warning: no upstream servers configured Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:06:26 localhost dnsmasq[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:26.715 265123 INFO neutron.agent.dhcp.agent [None req-2e47b0d3-c635-4de5-9362-924739681c02 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:24Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=bdfeb687-b9be-41a1-9990-15db976988d2, ip_allocation=immediate, mac_address=fa:16:3e:dc:25:b4, name=tempest-NetworksTestDHCPv6-1738840767, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=27, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['50fc47e8-c9d0-45e1-a5ff-3d725ccbecd1', 'ecc1998b-cb32-4980-94db-a8bdce306156'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:22Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1376, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:24Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:26 localhost nova_compute[284026]: 2025-11-27 10:06:26.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:26.856 2 INFO neutron.agent.securitygroups_rpc [None req-45899258-8640-4a42-858d-a09ee0efd1b3 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:26.966 265123 INFO neutron.agent.dhcp.agent [None req-efe5763b-a465-4365-a764-fae8841a0845 - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:26 localhost dnsmasq[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:26 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:26 localhost podman[318159]: 2025-11-27 10:06:26.988308845 +0000 UTC m=+0.063608274 container kill 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:06:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:27.102 2 INFO neutron.agent.securitygroups_rpc [None req-e0161151-71b5-41b1-83d0-d2db2c8660c6 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v254: 177 pgs: 177 active+clean; 192 MiB data, 838 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 3.3 KiB/s wr, 127 op/s Nov 27 05:06:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:27.334 2 INFO neutron.agent.securitygroups_rpc [None req-9812e936-3ddb-485e-8d39-5924c0a3152a d1dc452de09f4a5882dbe1b0b5007950 f53a81295fc84d788272f20e1bd6b608 - - default default] Security group member updated ['7fbf9cb2-f083-4cdd-8d2c-c899a55f43e5']#033[00m Nov 27 05:06:27 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:27.353 265123 INFO neutron.agent.dhcp.agent [None req-9f9cf88f-a448-4376-8807-c00826033456 - - - - - -] DHCP configuration for ports {'bdfeb687-b9be-41a1-9990-15db976988d2'} is completed#033[00m Nov 27 05:06:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:27.374 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '14'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:06:27 localhost dnsmasq[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:27 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:27 localhost dnsmasq-dhcp[318140]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:27 localhost podman[318197]: 2025-11-27 10:06:27.437576974 +0000 UTC m=+0.064132827 container kill 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e138 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:27.876 2 INFO neutron.agent.securitygroups_rpc [None req-45dc0ad0-42a7-442d-9ac1-1d399efd114b 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:28.059 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:27Z, description=, device_id=9ad9cc1f-1b9a-487f-a086-a258a5f09e3e, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e67e8d76-5707-41b0-8e9b-78cf55ed375b, ip_allocation=immediate, mac_address=fa:16:3e:8b:a0:bd, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1382, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:27Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:28 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:28.292 2 INFO neutron.agent.securitygroups_rpc [None req-9940c35a-886d-4d0a-a58c-a2b005058016 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:28 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:06:28 localhost podman[318238]: 2025-11-27 10:06:28.313048519 +0000 UTC m=+0.062699389 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:06:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:28.537 265123 INFO neutron.agent.dhcp.agent [None req-8fc03616-b4d9-470e-b99a-34f24fa4902e - - - - - -] DHCP configuration for ports {'e67e8d76-5707-41b0-8e9b-78cf55ed375b'} is completed#033[00m Nov 27 05:06:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:06:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:06:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e139 e139: 6 total, 6 up, 6 in Nov 27 05:06:29 localhost systemd[1]: tmp-crun.1m9grR.mount: Deactivated successfully. Nov 27 05:06:29 localhost podman[318259]: 2025-11-27 10:06:29.028751155 +0000 UTC m=+0.118918602 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_managed=true) Nov 27 05:06:29 localhost podman[318259]: 2025-11-27 10:06:29.065719187 +0000 UTC m=+0.155886664 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd) Nov 27 05:06:29 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:06:29 localhost dnsmasq[318140]: exiting on receipt of SIGTERM Nov 27 05:06:29 localhost podman[318305]: 2025-11-27 10:06:29.121811106 +0000 UTC m=+0.071053926 container kill 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:06:29 localhost systemd[1]: libpod-7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012.scope: Deactivated successfully. Nov 27 05:06:29 localhost podman[318258]: 2025-11-27 10:06:29.112507054 +0000 UTC m=+0.204886041 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v256: 177 pgs: 177 active+clean; 192 MiB data, 838 MiB used, 41 GiB / 42 GiB avail; 1.8 MiB/s rd, 3.2 KiB/s wr, 121 op/s Nov 27 05:06:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:29.194 2 INFO neutron.agent.securitygroups_rpc [None req-9194821e-c203-422b-8c27-064622152360 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:29 localhost podman[318326]: 2025-11-27 10:06:29.195308447 +0000 UTC m=+0.061800905 container died 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:06:29 localhost podman[318326]: 2025-11-27 10:06:29.231839957 +0000 UTC m=+0.098332365 container cleanup 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:06:29 localhost systemd[1]: libpod-conmon-7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012.scope: Deactivated successfully. Nov 27 05:06:29 localhost podman[318258]: 2025-11-27 10:06:29.24192844 +0000 UTC m=+0.334307407 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:29 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:06:29 localhost podman[318333]: 2025-11-27 10:06:29.325582225 +0000 UTC m=+0.180220272 container remove 7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.755 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.755 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.756 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:06:29 localhost nova_compute[284026]: 2025-11-27 10:06:29.757 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:06:30 localhost systemd[1]: var-lib-containers-storage-overlay-39b59d2ae05b8aa64d8fadb1b59468285a56578fe24c26a8833b91bc02998cf9-merged.mount: Deactivated successfully. Nov 27 05:06:30 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-7608725ff1967e3bfde8840af3ad4426b2662a3ef26ed07660ece45705242012-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:06:30 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1074309424' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.262 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.505s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.407 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.407 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.490 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.632 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.634 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11194MB free_disk=41.77423095703125GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.634 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.635 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.757 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.757 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.759 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.817 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:06:30 localhost podman[318429]: Nov 27 05:06:30 localhost podman[318429]: 2025-11-27 10:06:30.865283072 +0000 UTC m=+0.094854281 container create a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:06:30 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:30.897 2 INFO neutron.agent.securitygroups_rpc [None req-0f677e90-669f-49a3-b6ca-9b88517cadd6 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:30 localhost systemd[1]: Started libpod-conmon-a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44.scope. Nov 27 05:06:30 localhost podman[318429]: 2025-11-27 10:06:30.8179816 +0000 UTC m=+0.047552879 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:30 localhost systemd[1]: Started libcrun container. Nov 27 05:06:30 localhost nova_compute[284026]: 2025-11-27 10:06:30.946 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:30 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4510cbc4069440999ba92258bdef8d0a662289c8e5c87bca3cfa6293228ae57a/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:30 localhost podman[318429]: 2025-11-27 10:06:30.958515917 +0000 UTC m=+0.188087126 container init a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:06:30 localhost podman[318429]: 2025-11-27 10:06:30.974407658 +0000 UTC m=+0.203978867 container start a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:06:30 localhost dnsmasq[318467]: started, version 2.85 cachesize 150 Nov 27 05:06:30 localhost dnsmasq[318467]: DNS service limited to local subnets Nov 27 05:06:30 localhost dnsmasq[318467]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:30 localhost dnsmasq[318467]: warning: no upstream servers configured Nov 27 05:06:30 localhost dnsmasq-dhcp[318467]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:30 localhost dnsmasq[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:30 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:30 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:31 localhost systemd[1]: tmp-crun.x928GU.mount: Deactivated successfully. Nov 27 05:06:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v257: 177 pgs: 177 active+clean; 213 MiB data, 888 MiB used, 41 GiB / 42 GiB avail; 1.8 MiB/s rd, 3.4 MiB/s wr, 169 op/s Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.216 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:31.293 265123 INFO neutron.agent.dhcp.agent [None req-c78c5fd0-d5b7-46ea-b030-df96bd2d6eaa - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:06:31 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/924037983' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:06:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:06:31 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/924037983' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.331 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.514s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.337 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.373 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.375 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:06:31 localhost nova_compute[284026]: 2025-11-27 10:06:31.375 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.740s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:06:31 localhost dnsmasq[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:31 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:31 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:31 localhost podman[318489]: 2025-11-27 10:06:31.423618485 +0000 UTC m=+0.107002959 container kill a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:31 localhost podman[318516]: 2025-11-27 10:06:31.480283081 +0000 UTC m=+0.091580212 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:06:31 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:06:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:31 localhost systemd[1]: tmp-crun.2rTVUi.mount: Deactivated successfully. Nov 27 05:06:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:31.714 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '10', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '7', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:31.718 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:31.720 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:31.721 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:31.721 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a38eaeab-3c47-46dc-9ce3-a4c5b78ed69e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:31.759 265123 INFO neutron.agent.dhcp.agent [None req-8362c160-e4e6-4d6f-ac5d-3590e4f0c7c3 - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:31 localhost dnsmasq[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:31 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:31 localhost dnsmasq-dhcp[318467]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:31 localhost podman[318561]: 2025-11-27 10:06:31.894442849 +0000 UTC m=+0.048623348 container kill a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:31 localhost systemd-journald[49007]: Data hash table of /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal has a fill level at 75.0 (53725 of 71630 items, 25165824 file size, 468 bytes per hash table item), suggesting rotation. Nov 27 05:06:31 localhost systemd-journald[49007]: /run/log/journal/6d5b4ed5dba3c41afa293ccb7960bc9a/system.journal: Journal header limits reached or header out-of-date, rotating. Nov 27 05:06:31 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 05:06:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:32.000 2 INFO neutron.agent.securitygroups_rpc [None req-07e91ebc-2075-4220-93e6-037c08ad249c 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:32 localhost ovn_controller[156436]: 2025-11-27T10:06:32Z|00202|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:06:32 localhost rsyslogd[760]: imjournal: journal files changed, reloading... [v8.2102.0-111.el9 try https://www.rsyslog.com/e/0 ] Nov 27 05:06:32 localhost nova_compute[284026]: 2025-11-27 10:06:32.146 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 e140: 6 total, 6 up, 6 in Nov 27 05:06:32 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:32.251 265123 INFO neutron.agent.dhcp.agent [None req-1079e8de-0c4d-49a9-bfb7-7aa7bcab955d - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:32.587 2 INFO neutron.agent.securitygroups_rpc [None req-aa7a4d10-dbfe-4213-841e-d3b8a90395cb 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['83b550cc-ffe7-413e-86ab-de2527812517']#033[00m Nov 27 05:06:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v259: 177 pgs: 177 active+clean; 213 MiB data, 888 MiB used, 41 GiB / 42 GiB avail; 326 KiB/s rd, 3.4 MiB/s wr, 69 op/s Nov 27 05:06:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:33.355 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '11', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2'], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '10', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:33.357 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:33.360 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:33.361 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:33.362 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[bb960ba1-4cd4-453a-8892-827a82eb8e8c]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.375 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.376 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.376 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:06:33 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:33 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:33 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:33 localhost podman[318599]: 2025-11-27 10:06:33.457230481 +0000 UTC m=+0.066332947 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:06:33 localhost systemd[1]: tmp-crun.1hGFim.mount: Deactivated successfully. Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.546 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.547 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.547 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.548 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:06:33 localhost ovn_controller[156436]: 2025-11-27T10:06:33Z|00203|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:06:33 localhost dnsmasq[318467]: exiting on receipt of SIGTERM Nov 27 05:06:33 localhost podman[318637]: 2025-11-27 10:06:33.840234545 +0000 UTC m=+0.066962324 container kill a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:33 localhost systemd[1]: libpod-a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44.scope: Deactivated successfully. Nov 27 05:06:33 localhost nova_compute[284026]: 2025-11-27 10:06:33.843 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:33 localhost podman[318650]: 2025-11-27 10:06:33.908795732 +0000 UTC m=+0.054645931 container died a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:06:33 localhost podman[318650]: 2025-11-27 10:06:33.945092795 +0000 UTC m=+0.090942934 container cleanup a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:33 localhost systemd[1]: libpod-conmon-a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44.scope: Deactivated successfully. Nov 27 05:06:33 localhost podman[318652]: 2025-11-27 10:06:33.982578071 +0000 UTC m=+0.120277379 container remove a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:06:34 localhost systemd[1]: var-lib-containers-storage-overlay-4510cbc4069440999ba92258bdef8d0a662289c8e5c87bca3cfa6293228ae57a-merged.mount: Deactivated successfully. Nov 27 05:06:34 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-a08354f0e1e00047352d21370e23d79069893087aa82741d6b7849d78e1d5e44-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:34 localhost nova_compute[284026]: 2025-11-27 10:06:34.573 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:06:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:34.584 2 INFO neutron.agent.securitygroups_rpc [None req-bddffe80-57e7-45f5-96b9-6b1bb0a60092 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['3908970f-ca93-4993-9b9a-30f6a70ca63d']#033[00m Nov 27 05:06:34 localhost nova_compute[284026]: 2025-11-27 10:06:34.609 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:06:34 localhost nova_compute[284026]: 2025-11-27 10:06:34.610 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:06:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:34.966 2 INFO neutron.agent.securitygroups_rpc [None req-2d281f12-1bc7-4cda-8986-039ab225e2d7 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v260: 177 pgs: 177 active+clean; 213 MiB data, 888 MiB used, 41 GiB / 42 GiB avail; 296 KiB/s rd, 3.1 MiB/s wr, 63 op/s Nov 27 05:06:35 localhost nova_compute[284026]: 2025-11-27 10:06:35.530 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:35 localhost podman[318730]: Nov 27 05:06:35 localhost podman[318730]: 2025-11-27 10:06:35.704080352 +0000 UTC m=+0.089758683 container create cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:35 localhost podman[318730]: 2025-11-27 10:06:35.663668887 +0000 UTC m=+0.049347238 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:35 localhost podman[318761]: 2025-11-27 10:06:35.795448747 +0000 UTC m=+0.060070008 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:35 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:06:35 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:35 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:35 localhost systemd[1]: Started libpod-conmon-cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107.scope. Nov 27 05:06:35 localhost systemd[1]: Started libcrun container. Nov 27 05:06:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/2ef43fc6433b4d617a4a3b2f54818d7c9547fbe86646adcc3bdd5f4ef9f4e273/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:35 localhost podman[318730]: 2025-11-27 10:06:35.823434635 +0000 UTC m=+0.209112966 container init cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:06:35 localhost podman[318730]: 2025-11-27 10:06:35.83136788 +0000 UTC m=+0.217046201 container start cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:35 localhost dnsmasq[318780]: started, version 2.85 cachesize 150 Nov 27 05:06:35 localhost dnsmasq[318780]: DNS service limited to local subnets Nov 27 05:06:35 localhost dnsmasq[318780]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:35 localhost dnsmasq[318780]: warning: no upstream servers configured Nov 27 05:06:35 localhost dnsmasq-dhcp[318780]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:35 localhost dnsmasq[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:35 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:35 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:35 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:35.884 265123 INFO neutron.agent.dhcp.agent [None req-aec03bce-53a8-49d1-a6d1-cbff2edeb4c6 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:34Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=663acd1f-822a-45bc-bfc9-d62698fe9f8d, ip_allocation=immediate, mac_address=fa:16:3e:35:fb:4d, name=tempest-NetworksTestDHCPv6-193766098, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=31, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['d812cf57-e5ee-42d9-845c-b903cb8e03db', 'f765ba6f-9293-4f19-bdc5-69e06900b744'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:31Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1398, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:34Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:35 localhost ovn_controller[156436]: 2025-11-27T10:06:35Z|00204|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:06:35 localhost nova_compute[284026]: 2025-11-27 10:06:35.916 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:35 localhost nova_compute[284026]: 2025-11-27 10:06:35.946 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:35 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:35.962 2 INFO neutron.agent.securitygroups_rpc [None req-36d45ede-3011-409b-97a8-6dde0fd3c8f9 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['7d38ec37-83a0-43d5-8f3a-eaa2b0e9c8e9']#033[00m Nov 27 05:06:36 localhost podman[318803]: 2025-11-27 10:06:36.133421892 +0000 UTC m=+0.053598514 container kill cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:06:36 localhost dnsmasq[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:06:36 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:36 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:36.154 265123 INFO neutron.agent.dhcp.agent [None req-2f02021e-7861-46cc-893a-d514fd0b346f - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:36 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:36.212 2 INFO neutron.agent.securitygroups_rpc [None req-d7aa1053-dc64-4891-bed0-f68045e479c0 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:36 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:36.228 2 INFO neutron.agent.securitygroups_rpc [None req-9c1db2e1-3030-4837-96dc-4b1a3c5e931b 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['7d38ec37-83a0-43d5-8f3a-eaa2b0e9c8e9']#033[00m Nov 27 05:06:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:36.379 265123 INFO neutron.agent.dhcp.agent [None req-65ecbfb5-1b3a-418e-b46e-ce8dceb251f6 - - - - - -] DHCP configuration for ports {'663acd1f-822a-45bc-bfc9-d62698fe9f8d'} is completed#033[00m Nov 27 05:06:36 localhost dnsmasq[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:36 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:36 localhost dnsmasq-dhcp[318780]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:36 localhost podman[318843]: 2025-11-27 10:06:36.521938915 +0000 UTC m=+0.054892577 container kill cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:06:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:36.653 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:36Z, description=, device_id=1cd6e997-1bc3-4f4f-869f-70b315f30e8d, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=13a570ad-4e85-4cf4-ab56-0d338f91c65a, ip_allocation=immediate, mac_address=fa:16:3e:87:ac:28, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1418, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:36Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:36 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:36 localhost podman[318883]: 2025-11-27 10:06:36.894580469 +0000 UTC m=+0.055941186 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:06:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:37 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:37.159 265123 INFO neutron.agent.dhcp.agent [None req-9dc25ebe-cad3-4a34-adce-5dea042c1884 - - - - - -] DHCP configuration for ports {'13a570ad-4e85-4cf4-ab56-0d338f91c65a'} is completed#033[00m Nov 27 05:06:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v261: 177 pgs: 177 active+clean; 225 MiB data, 953 MiB used, 41 GiB / 42 GiB avail; 515 KiB/s rd, 3.1 MiB/s wr, 129 op/s Nov 27 05:06:37 localhost dnsmasq[318780]: exiting on receipt of SIGTERM Nov 27 05:06:37 localhost podman[318921]: 2025-11-27 10:06:37.62668646 +0000 UTC m=+0.061669261 container kill cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:37 localhost systemd[1]: libpod-cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107.scope: Deactivated successfully. Nov 27 05:06:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:06:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:06:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:06:37 localhost podman[318936]: 2025-11-27 10:06:37.718668141 +0000 UTC m=+0.069556344 container died cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:06:37 localhost systemd[1]: tmp-crun.9O59oN.mount: Deactivated successfully. Nov 27 05:06:37 localhost systemd[1]: var-lib-containers-storage-overlay-2ef43fc6433b4d617a4a3b2f54818d7c9547fbe86646adcc3bdd5f4ef9f4e273-merged.mount: Deactivated successfully. Nov 27 05:06:37 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:37 localhost podman[318936]: 2025-11-27 10:06:37.759809936 +0000 UTC m=+0.110698099 container remove cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:06:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:37 localhost podman[318949]: 2025-11-27 10:06:37.849232668 +0000 UTC m=+0.187457699 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, vendor=Red Hat, Inc., container_name=openstack_network_exporter, io.openshift.expose-services=, architecture=x86_64, io.buildah.version=1.33.7, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, maintainer=Red Hat, Inc., managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, version=9.6, name=ubi9-minimal, distribution-scope=public, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, release=1755695350) Nov 27 05:06:37 localhost podman[318949]: 2025-11-27 10:06:37.89137825 +0000 UTC m=+0.229603271 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, com.redhat.component=ubi9-minimal-container, managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, maintainer=Red Hat, Inc., architecture=x86_64, vendor=Red Hat, Inc., container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public) Nov 27 05:06:37 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:06:37 localhost podman[318938]: 2025-11-27 10:06:37.907506237 +0000 UTC m=+0.254752512 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:37 localhost systemd[1]: libpod-conmon-cdf89fa301a55b40833665c9fcdf41e6cf65b869f2797cbc0d474e9de0950107.scope: Deactivated successfully. Nov 27 05:06:37 localhost podman[318938]: 2025-11-27 10:06:37.922198675 +0000 UTC m=+0.269444960 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2) Nov 27 05:06:37 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:06:38 localhost podman[318943]: 2025-11-27 10:06:37.999987522 +0000 UTC m=+0.343362912 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:06:38 localhost podman[318943]: 2025-11-27 10:06:38.013421916 +0000 UTC m=+0.356797316 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:06:38 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:06:38 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:38.274 2 INFO neutron.agent.securitygroups_rpc [None req-8ebd1b54-fe57-4526-aa07-7989bfce1188 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['c42fdf9e-5fa3-4dc6-800a-b33ac5e6e737']#033[00m Nov 27 05:06:38 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:38.520 2 INFO neutron.agent.securitygroups_rpc [None req-cf737238-cc40-4f1e-b57d-27ed0c6cb6b4 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['c42fdf9e-5fa3-4dc6-800a-b33ac5e6e737']#033[00m Nov 27 05:06:38 localhost podman[319073]: Nov 27 05:06:38 localhost podman[319073]: 2025-11-27 10:06:38.665059957 +0000 UTC m=+0.083733630 container create b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:38 localhost systemd[1]: Started libpod-conmon-b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7.scope. Nov 27 05:06:38 localhost systemd[1]: Started libcrun container. Nov 27 05:06:38 localhost podman[319073]: 2025-11-27 10:06:38.623789399 +0000 UTC m=+0.042463122 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:38 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5256266789cd4e3a3c1f0d84587cf1f1d8a3faab63ed808c66992386b363142e/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:38 localhost podman[319073]: 2025-11-27 10:06:38.733397688 +0000 UTC m=+0.152071341 container init b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:06:38 localhost podman[319073]: 2025-11-27 10:06:38.742675989 +0000 UTC m=+0.161349642 container start b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:38 localhost dnsmasq[319091]: started, version 2.85 cachesize 150 Nov 27 05:06:38 localhost dnsmasq[319091]: DNS service limited to local subnets Nov 27 05:06:38 localhost dnsmasq[319091]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:38 localhost dnsmasq[319091]: warning: no upstream servers configured Nov 27 05:06:38 localhost dnsmasq-dhcp[319091]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:38 localhost dnsmasq[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:38 localhost dnsmasq-dhcp[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:38 localhost dnsmasq-dhcp[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:38 localhost podman[242678]: time="2025-11-27T10:06:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:06:38 localhost podman[242678]: @ - - [27/Nov/2025:10:06:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159336 "" "Go-http-client/1.1" Nov 27 05:06:38 localhost podman[242678]: @ - - [27/Nov/2025:10:06:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20214 "" "Go-http-client/1.1" Nov 27 05:06:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v262: 177 pgs: 177 active+clean; 225 MiB data, 953 MiB used, 41 GiB / 42 GiB avail; 424 KiB/s rd, 2.6 MiB/s wr, 106 op/s Nov 27 05:06:39 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:39.184 265123 INFO neutron.agent.dhcp.agent [None req-660c92b2-f221-4dd1-aa3e-09d3c06adb2d - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:39.322 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '14', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '11', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:39 localhost podman[319109]: 2025-11-27 10:06:39.324608083 +0000 UTC m=+0.061664262 container kill b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:06:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:39.325 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:39 localhost dnsmasq[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:39 localhost dnsmasq-dhcp[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:39 localhost dnsmasq-dhcp[319091]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:39.327 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:39.328 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:39 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:39.329 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[993e8308-bccd-4bb8-a28f-b29d4eb5bb89]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:39 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:39.643 265123 INFO neutron.agent.dhcp.agent [None req-f2bf52f2-66a2-4ca4-8421-7fe5df3ed1db - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:39 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:39.776 2 INFO neutron.agent.securitygroups_rpc [None req-b8017921-6336-4633-8e44-a6c375205fc8 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:40.209 2 INFO neutron.agent.securitygroups_rpc [None req-aac61094-20cc-43e6-8ee9-5c333fe4e7fb 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:40 localhost nova_compute[284026]: 2025-11-27 10:06:40.555 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:40 localhost nova_compute[284026]: 2025-11-27 10:06:40.951 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v263: 177 pgs: 177 active+clean; 225 MiB data, 954 MiB used, 41 GiB / 42 GiB avail; 188 KiB/s rd, 130 KiB/s wr, 56 op/s Nov 27 05:06:41 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:41.426 2 INFO neutron.agent.securitygroups_rpc [None req-b4b3ed4e-3893-4b2d-8bae-1ea6f768e4e8 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:42.209 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '15', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2'], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '14', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:42.210 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:42.212 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 76d19211-53be-46fd-878a-569122192197 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:42.212 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:42 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:42.213 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[129fcd36-bf15-4983-b42a-6b40756d2be7]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:42 localhost podman[319146]: 2025-11-27 10:06:42.309459112 +0000 UTC m=+0.061339482 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125) Nov 27 05:06:42 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:06:42 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:42 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:42.329 2 INFO neutron.agent.securitygroups_rpc [None req-31d3cb57-60e7-4f62-9d37-7e74d25c4688 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:42 localhost podman[319184]: 2025-11-27 10:06:42.714104904 +0000 UTC m=+0.056348327 container kill b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:06:42 localhost dnsmasq[319091]: exiting on receipt of SIGTERM Nov 27 05:06:42 localhost systemd[1]: libpod-b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7.scope: Deactivated successfully. Nov 27 05:06:42 localhost podman[319196]: 2025-11-27 10:06:42.777351107 +0000 UTC m=+0.051508806 container died b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:06:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:42 localhost podman[319196]: 2025-11-27 10:06:42.810984868 +0000 UTC m=+0.085142527 container cleanup b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:06:42 localhost systemd[1]: libpod-conmon-b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7.scope: Deactivated successfully. Nov 27 05:06:42 localhost podman[319198]: 2025-11-27 10:06:42.866914623 +0000 UTC m=+0.132099709 container remove b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:06:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v264: 177 pgs: 177 active+clean; 225 MiB data, 954 MiB used, 41 GiB / 42 GiB avail; 172 KiB/s rd, 119 KiB/s wr, 51 op/s Nov 27 05:06:43 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:43.299 2 INFO neutron.agent.securitygroups_rpc [None req-73f9ea6b-1fd8-4282-b6e1-74caee13be48 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:43 localhost systemd[1]: var-lib-containers-storage-overlay-5256266789cd4e3a3c1f0d84587cf1f1d8a3faab63ed808c66992386b363142e-merged.mount: Deactivated successfully. Nov 27 05:06:43 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-b26f7442c8db806ec51efbd791c563bfa1714a46744fe8b7e8ba8c0e3a0775f7-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:43.575 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:06:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:43.576 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:06:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:43.578 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:06:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:06:44 Nov 27 05:06:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:06:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:06:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['manila_data', 'volumes', 'images', '.mgr', 'backups', 'manila_metadata', 'vms'] Nov 27 05:06:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:06:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:44.469 2 INFO neutron.agent.securitygroups_rpc [None req-c9112313-745d-4695-a899-552e4b3a5652 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['38b5ba25-00fc-467e-b329-d63c8ba8f13d']#033[00m Nov 27 05:06:44 localhost podman[319278]: Nov 27 05:06:44 localhost podman[319278]: 2025-11-27 10:06:44.492316091 +0000 UTC m=+0.093503763 container create 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:44 localhost systemd[1]: Started libpod-conmon-70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79.scope. Nov 27 05:06:44 localhost systemd[1]: Started libcrun container. Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:44 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5856e3deed28b772e4ea65d15522b0431c5cf7a1f7bd48408e8ee8c3e0f3ec7c/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:44 localhost podman[319278]: 2025-11-27 10:06:44.443952801 +0000 UTC m=+0.045140503 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:44 localhost podman[319278]: 2025-11-27 10:06:44.55318202 +0000 UTC m=+0.154369692 container init 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:06:44 localhost podman[319278]: 2025-11-27 10:06:44.561772233 +0000 UTC m=+0.162959925 container start 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:06:44 localhost dnsmasq[319296]: started, version 2.85 cachesize 150 Nov 27 05:06:44 localhost dnsmasq[319296]: DNS service limited to local subnets Nov 27 05:06:44 localhost dnsmasq[319296]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:44 localhost dnsmasq[319296]: warning: no upstream servers configured Nov 27 05:06:44 localhost dnsmasq-dhcp[319296]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:44 localhost dnsmasq-dhcp[319296]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:06:44 localhost dnsmasq[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:44 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:44 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:06:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0065810279697096595 of space, bias 1.0, pg target 1.316205593941932 quantized to 32 (current 32) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 1.0905220547180346e-06 of space, bias 1.0, pg target 0.00021701388888888888 quantized to 32 (current 32) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8555772569444443 quantized to 32 (current 32) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:06:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.0021628687418574354 quantized to 16 (current 16) Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:06:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:06:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:44.723 2 INFO neutron.agent.securitygroups_rpc [req-2e45f1b6-c9b8-4ae7-aa9a-83011b4f6cd7 req-da0d5d7c-fa63-48a1-aee8-48a7ef624acf 88f0516a86664a828ee0e3a2264f4672 f0753b0eb35c49d1852f196ec33a2877 - - default default] Security group member updated ['677d6692-9431-43e9-bf62-1d9fe6051ff5']#033[00m Nov 27 05:06:44 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:44.724 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:44Z, description=, device_id=10909cf3-7785-41ad-875d-63cca4f4289e, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=05a1f0e6-aac9-477d-9ea0-5761f01ac810, ip_allocation=immediate, mac_address=fa:16:3e:a6:e5:2b, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1474, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:44Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:44.820 2 INFO neutron.agent.securitygroups_rpc [None req-203e65ad-fe73-4ddd-9c11-cc89c49d0f77 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:44 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:44.929 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:43Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=4dd59de4-5dbf-4d89-a0c5-5bdf63daa062, ip_allocation=immediate, mac_address=fa:16:3e:e8:69:1f, name=tempest-NetworksTestDHCPv6-1791071727, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=35, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['50ddb9de-4f55-44de-9f6e-a8d3975638f9', 'f8419069-5059-4ae0-86a6-a26544ef8044'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:39Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1472, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:44Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:44 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:44 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:44 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:44 localhost podman[319314]: 2025-11-27 10:06:44.974918064 +0000 UTC m=+0.059516274 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:45.077 265123 INFO neutron.agent.dhcp.agent [None req-dcef286f-5758-462a-845d-ce73a0f23b7d - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:45.127 265123 INFO neutron.agent.dhcp.agent [None req-bec41a3f-143a-4126-942a-5c632a72fafa - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:44Z, description=, device_id=1cbfa3df-cc20-43b1-aa31-4e61ba5b27bc, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=71977935-e65f-4c22-9a49-986e71253cd9, ip_allocation=immediate, mac_address=fa:16:3e:5e:15:68, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1473, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:44Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v265: 177 pgs: 177 active+clean; 225 MiB data, 954 MiB used, 41 GiB / 42 GiB avail; 157 KiB/s rd, 109 KiB/s wr, 47 op/s Nov 27 05:06:45 localhost dnsmasq[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:06:45 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:45 localhost podman[319353]: 2025-11-27 10:06:45.199406295 +0000 UTC m=+0.058205338 container kill 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:45 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:45.295 265123 INFO neutron.agent.dhcp.agent [None req-c3434e9c-f0a1-46cb-942d-1e300d2735e0 - - - - - -] DHCP configuration for ports {'05a1f0e6-aac9-477d-9ea0-5761f01ac810'} is completed#033[00m Nov 27 05:06:45 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:06:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:45 localhost podman[319386]: 2025-11-27 10:06:45.32768279 +0000 UTC m=+0.058919547 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:06:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:06:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:06:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:45.471 265123 INFO neutron.agent.dhcp.agent [None req-cdbb499d-a9c8-4a4b-a091-88bb0042e439 - - - - - -] DHCP configuration for ports {'4dd59de4-5dbf-4d89-a0c5-5bdf63daa062'} is completed#033[00m Nov 27 05:06:45 localhost podman[319407]: 2025-11-27 10:06:45.481231548 +0000 UTC m=+0.076133013 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:06:45 localhost podman[319407]: 2025-11-27 10:06:45.517048129 +0000 UTC m=+0.111949624 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:06:45 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:06:45 localhost podman[319409]: 2025-11-27 10:06:45.53074624 +0000 UTC m=+0.122744026 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:06:45 localhost nova_compute[284026]: 2025-11-27 10:06:45.557 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:45 localhost podman[319409]: 2025-11-27 10:06:45.56694613 +0000 UTC m=+0.158943947 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:06:45 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:06:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:45.654 265123 INFO neutron.agent.dhcp.agent [None req-b7fb7536-965d-413d-abb8-57380db72146 - - - - - -] DHCP configuration for ports {'71977935-e65f-4c22-9a49-986e71253cd9'} is completed#033[00m Nov 27 05:06:45 localhost nova_compute[284026]: 2025-11-27 10:06:45.953 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:46 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:46.568 2 INFO neutron.agent.securitygroups_rpc [None req-be1c23d4-4a4b-4d76-b659-368a4b1f90c3 3339fb7d6c9a40499e41a69207bb1634 329c45e899cd4311b5fd9be149f66671 - - default default] Security group rule updated ['b24b37ee-dfad-4eed-81ab-5223bdc3b25d']#033[00m Nov 27 05:06:46 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:46.956 2 INFO neutron.agent.securitygroups_rpc [None req-fd7567d0-1104-4f10-ae49-d5a4c374b40b 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v266: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 176 KiB/s rd, 110 KiB/s wr, 74 op/s Nov 27 05:06:47 localhost nova_compute[284026]: 2025-11-27 10:06:47.178 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:47 localhost dnsmasq[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:47 localhost podman[319478]: 2025-11-27 10:06:47.309644756 +0000 UTC m=+0.056429100 container kill 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:06:47 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:47 localhost dnsmasq-dhcp[319296]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:48 localhost dnsmasq[319296]: exiting on receipt of SIGTERM Nov 27 05:06:48 localhost podman[319519]: 2025-11-27 10:06:48.282174999 +0000 UTC m=+0.060876569 container kill 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:06:48 localhost systemd[1]: libpod-70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79.scope: Deactivated successfully. Nov 27 05:06:48 localhost podman[319532]: 2025-11-27 10:06:48.355129135 +0000 UTC m=+0.055996468 container died 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:48 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:48 localhost systemd[1]: var-lib-containers-storage-overlay-5856e3deed28b772e4ea65d15522b0431c5cf7a1f7bd48408e8ee8c3e0f3ec7c-merged.mount: Deactivated successfully. Nov 27 05:06:48 localhost podman[319532]: 2025-11-27 10:06:48.389488906 +0000 UTC m=+0.090356199 container cleanup 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:06:48 localhost systemd[1]: libpod-conmon-70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79.scope: Deactivated successfully. Nov 27 05:06:48 localhost podman[319534]: 2025-11-27 10:06:48.441266838 +0000 UTC m=+0.134813772 container remove 70326525e190f6fe9ed53425b1315531a809f091a64e23e0bc0c2009b3376c79 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:06:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:06:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3578463773' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:06:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:06:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3578463773' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:06:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v267: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 14 KiB/s wr, 28 op/s Nov 27 05:06:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:49.470 2 INFO neutron.agent.securitygroups_rpc [None req-1713e59b-d344-4ebf-b7de-172e143600dc 9cd13abea9f04c7c9acaacd0c9346e8c 5711c25eda684e6595773005b3d0d3af - - default default] Security group member updated ['3efaea02-6ea0-4b8a-abcb-9a7f7662a409']#033[00m Nov 27 05:06:49 localhost podman[319610]: Nov 27 05:06:49 localhost podman[319610]: 2025-11-27 10:06:49.49457164 +0000 UTC m=+0.090799920 container create 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:06:49 localhost systemd[1]: Started libpod-conmon-1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d.scope. Nov 27 05:06:49 localhost podman[319610]: 2025-11-27 10:06:49.449955091 +0000 UTC m=+0.046183411 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:49 localhost systemd[1]: Started libcrun container. Nov 27 05:06:49 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/d4e22529c61e517150b768bc9b57bedce5e1d834b3b4202ae0a3de4fcea9751f/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:49 localhost podman[319610]: 2025-11-27 10:06:49.570435555 +0000 UTC m=+0.166663835 container init 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:06:49 localhost podman[319610]: 2025-11-27 10:06:49.579026557 +0000 UTC m=+0.175254837 container start 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:06:49 localhost dnsmasq[319628]: started, version 2.85 cachesize 150 Nov 27 05:06:49 localhost dnsmasq[319628]: DNS service limited to local subnets Nov 27 05:06:49 localhost dnsmasq[319628]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:49 localhost dnsmasq[319628]: warning: no upstream servers configured Nov 27 05:06:49 localhost dnsmasq-dhcp[319628]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:06:49 localhost dnsmasq[319628]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:49 localhost dnsmasq-dhcp[319628]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:06:49 localhost dnsmasq-dhcp[319628]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:06:49 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:49.665 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '17', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 10.100.0.2 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '10.100.0.2/28 2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '15', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:49 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:49.668 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:06:49 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:49.670 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:06:49 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:49.671 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[f86efada-41ee-4d78-858b-72f04ac01531]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:49 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:49 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:49 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:49 localhost podman[319647]: 2025-11-27 10:06:49.877749529 +0000 UTC m=+0.061307231 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:06:50 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:50.030 265123 INFO neutron.agent.dhcp.agent [None req-25ba4c99-4f44-4f10-b395-70819747d238 - - - - - -] DHCP configuration for ports {'226ebddc-31ad-4e27-8efb-98e9f84cfce2', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:50 localhost dnsmasq[319628]: exiting on receipt of SIGTERM Nov 27 05:06:50 localhost podman[319678]: 2025-11-27 10:06:50.038676248 +0000 UTC m=+0.072500285 container kill 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:06:50 localhost systemd[1]: libpod-1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d.scope: Deactivated successfully. Nov 27 05:06:50 localhost podman[319694]: 2025-11-27 10:06:50.111922952 +0000 UTC m=+0.058541266 container died 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:50 localhost podman[319694]: 2025-11-27 10:06:50.139932851 +0000 UTC m=+0.086551115 container cleanup 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:06:50 localhost systemd[1]: libpod-conmon-1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d.scope: Deactivated successfully. Nov 27 05:06:50 localhost podman[319696]: 2025-11-27 10:06:50.189062312 +0000 UTC m=+0.129006386 container remove 1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:06:50 localhost nova_compute[284026]: 2025-11-27 10:06:50.201 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:50 localhost ovn_controller[156436]: 2025-11-27T10:06:50Z|00205|binding|INFO|Releasing lport 226ebddc-31ad-4e27-8efb-98e9f84cfce2 from this chassis (sb_readonly=0) Nov 27 05:06:50 localhost ovn_controller[156436]: 2025-11-27T10:06:50Z|00206|binding|INFO|Setting lport 226ebddc-31ad-4e27-8efb-98e9f84cfce2 down in Southbound Nov 27 05:06:50 localhost kernel: device tap226ebddc-31 left promiscuous mode Nov 27 05:06:50 localhost nova_compute[284026]: 2025-11-27 10:06:50.221 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:50.362 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.3/28 2001:db8::f816:3eff:fe98:4365/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '12', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=226ebddc-31ad-4e27-8efb-98e9f84cfce2) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:50.364 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 226ebddc-31ad-4e27-8efb-98e9f84cfce2 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:06:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:50.366 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1ba86240-13a7-470b-9c24-36a566ecc5cd or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:06:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:50.367 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ff87a3c3-2639-420e-8b4c-97942776b21d]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:50 localhost systemd[1]: var-lib-containers-storage-overlay-d4e22529c61e517150b768bc9b57bedce5e1d834b3b4202ae0a3de4fcea9751f-merged.mount: Deactivated successfully. Nov 27 05:06:50 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-1154f13beb63bbdc0fb8e363e15faa0fbf0039f2925da80de429838a6095ae3d-userdata-shm.mount: Deactivated successfully. Nov 27 05:06:50 localhost nova_compute[284026]: 2025-11-27 10:06:50.559 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:50 localhost nova_compute[284026]: 2025-11-27 10:06:50.958 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v268: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 14 KiB/s wr, 28 op/s Nov 27 05:06:52 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:06:52 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:52.399 265123 INFO neutron.agent.dhcp.agent [None req-02ad5ede-5548-4d0b-b136-535b1d2615b0 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:06:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:53 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:53.127 2 INFO neutron.agent.securitygroups_rpc [None req-e557b947-95db-413e-8656-53e54830d8f0 b4d41172b1f240759776aff8a68ef7d7 f4801ccffc5e49c4b1f84c728382de18 - - default default] Security group member updated ['32da611f-940b-4d76-9ca5-88abf5278da1']#033[00m Nov 27 05:06:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v269: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 1.2 KiB/s wr, 27 op/s Nov 27 05:06:53 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:53.967 2 INFO neutron.agent.securitygroups_rpc [None req-31916d5d-8fd5-4596-a3ba-9bccf6707b17 b4d41172b1f240759776aff8a68ef7d7 f4801ccffc5e49c4b1f84c728382de18 - - default default] Security group member updated ['32da611f-940b-4d76-9ca5-88abf5278da1']#033[00m Nov 27 05:06:54 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:54.825 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:54Z, description=, device_id=ecb7311b-c513-4976-a869-2ba2a257a6a7, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=a3086065-afe3-4ac8-bdc9-b0957b54f73d, ip_allocation=immediate, mac_address=fa:16:3e:cb:ee:a7, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1515, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:06:54Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:06:55 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:06:55 localhost podman[319741]: 2025-11-27 10:06:55.052738747 +0000 UTC m=+0.059991636 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:55 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:55 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v270: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 1.2 KiB/s wr, 27 op/s Nov 27 05:06:55 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:55.333 265123 INFO neutron.agent.dhcp.agent [None req-9a085dd0-f36b-4e7f-ba59-f36d72d87a9c - - - - - -] DHCP configuration for ports {'a3086065-afe3-4ac8-bdc9-b0957b54f73d'} is completed#033[00m Nov 27 05:06:55 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:55.458 265123 INFO neutron.agent.linux.ip_lib [None req-d7688ac4-082b-430c-9ad6-85b58d0d1517 - - - - - -] Device tap723a44c5-55 cannot be used as it has no MAC address#033[00m Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.485 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost kernel: device tap723a44c5-55 entered promiscuous mode Nov 27 05:06:55 localhost NetworkManager[5971]: [1764238015.4952] manager: (tap723a44c5-55): new Generic device (/org/freedesktop/NetworkManager/Devices/38) Nov 27 05:06:55 localhost ovn_controller[156436]: 2025-11-27T10:06:55Z|00207|binding|INFO|Claiming lport 723a44c5-5507-4969-8f33-21dec93014d9 for this chassis. Nov 27 05:06:55 localhost ovn_controller[156436]: 2025-11-27T10:06:55Z|00208|binding|INFO|723a44c5-5507-4969-8f33-21dec93014d9: Claiming unknown Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.498 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost systemd-udevd[319773]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:06:55 localhost ovn_controller[156436]: 2025-11-27T10:06:55Z|00209|binding|INFO|Setting lport 723a44c5-5507-4969-8f33-21dec93014d9 up in Southbound Nov 27 05:06:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:55.504 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe5a:b6ef/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=723a44c5-5507-4969-8f33-21dec93014d9) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.504 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost ovn_controller[156436]: 2025-11-27T10:06:55Z|00210|binding|INFO|Setting lport 723a44c5-5507-4969-8f33-21dec93014d9 ovn-installed in OVS Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.505 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:55.506 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 723a44c5-5507-4969-8f33-21dec93014d9 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:06:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:55.509 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 142f9962-ce03-47a6-a6a4-85e807b0407f IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:06:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:55.509 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:06:55 localhost ovn_metadata_agent[162087]: 2025-11-27 10:06:55.510 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[80b999c8-614f-4597-84d1-148797a6cbd9]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.511 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.538 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.561 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost journal[232028]: ethtool ioctl error on tap723a44c5-55: No such device Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.579 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.602 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:55 localhost openstack_network_exporter[244641]: ERROR 10:06:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:06:55 localhost openstack_network_exporter[244641]: ERROR 10:06:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:06:55 localhost openstack_network_exporter[244641]: ERROR 10:06:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:06:55 localhost openstack_network_exporter[244641]: ERROR 10:06:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:06:55 localhost openstack_network_exporter[244641]: Nov 27 05:06:55 localhost openstack_network_exporter[244641]: ERROR 10:06:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:06:55 localhost openstack_network_exporter[244641]: Nov 27 05:06:55 localhost nova_compute[284026]: 2025-11-27 10:06:55.996 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:56 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:56.405 2 INFO neutron.agent.securitygroups_rpc [None req-3570a3de-2d3c-449f-91ba-4370ee8c0ad0 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:56 localhost podman[319844]: Nov 27 05:06:56 localhost podman[319844]: 2025-11-27 10:06:56.45091422 +0000 UTC m=+0.095105797 container create 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:06:56 localhost systemd[1]: Started libpod-conmon-0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245.scope. Nov 27 05:06:56 localhost podman[319844]: 2025-11-27 10:06:56.408211253 +0000 UTC m=+0.052402910 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:06:56 localhost systemd[1]: Started libcrun container. Nov 27 05:06:56 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/53138c1a564f90e694ccf06cdefd8f57fbb42d0cdb23d5bde3928385505d84e2/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:06:56 localhost podman[319844]: 2025-11-27 10:06:56.517718909 +0000 UTC m=+0.161910516 container init 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:06:56 localhost podman[319844]: 2025-11-27 10:06:56.526012794 +0000 UTC m=+0.170204391 container start 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:06:56 localhost dnsmasq[319862]: started, version 2.85 cachesize 150 Nov 27 05:06:56 localhost dnsmasq[319862]: DNS service limited to local subnets Nov 27 05:06:56 localhost dnsmasq[319862]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:06:56 localhost dnsmasq[319862]: warning: no upstream servers configured Nov 27 05:06:56 localhost dnsmasq[319862]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:06:56 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:56.583 265123 INFO neutron.agent.dhcp.agent [None req-d7688ac4-082b-430c-9ad6-85b58d0d1517 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:55Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3b8afba1-90ed-4f11-bb3e-01ed1517f84b, ip_allocation=immediate, mac_address=fa:16:3e:43:e1:77, name=tempest-NetworksTestDHCPv6-1785071455, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=38, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['79ca95c0-83dd-4336-b3f3-6bbe6f25779e'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:52Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1519, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:06:55Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:06:56 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:56.693 265123 INFO neutron.agent.dhcp.agent [None req-78073a2e-77cd-4552-aa11-86cf3d37c595 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:06:56 localhost dnsmasq[319862]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:06:56 localhost podman[319881]: 2025-11-27 10:06:56.775964064 +0000 UTC m=+0.057496417 container kill 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:06:57 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:57.112 2 INFO neutron.agent.securitygroups_rpc [None req-2e9d157c-051e-41fc-9dcb-83943438bbb6 8c95b3c0401d4587a75defa3e8e50842 54da21f9a9d44c6ca40cac98e791742b - - default default] Security group rule updated ['779669d2-dfb9-40c3-8310-bab141bbc480']#033[00m Nov 27 05:06:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:06:57.124 265123 INFO neutron.agent.dhcp.agent [None req-bf4174c0-e8a3-4d5d-85cb-29e56908e1f4 - - - - - -] DHCP configuration for ports {'3b8afba1-90ed-4f11-bb3e-01ed1517f84b'} is completed#033[00m Nov 27 05:06:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v271: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail; 19 KiB/s rd, 1.2 KiB/s wr, 27 op/s Nov 27 05:06:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:06:58 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:06:58 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:58 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:58 localhost podman[319917]: 2025-11-27 10:06:58.031583047 +0000 UTC m=+0.050751897 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:06:58 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:06:58 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:06:58 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:06:58 localhost podman[319954]: 2025-11-27 10:06:58.904608634 +0000 UTC m=+0.060783888 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:06:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v272: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:06:59 localhost ovn_controller[156436]: 2025-11-27T10:06:59Z|00211|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:06:59 localhost nova_compute[284026]: 2025-11-27 10:06:59.310 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:59 localhost neutron_sriov_agent[258162]: 2025-11-27 10:06:59.722 2 INFO neutron.agent.securitygroups_rpc [None req-a5973ed1-bb5d-4ae0-8517-59171ff0c4fd 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:06:59 localhost nova_compute[284026]: 2025-11-27 10:06:59.762 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:06:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:06:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:06:59 localhost systemd[1]: tmp-crun.QFuCwY.mount: Deactivated successfully. Nov 27 05:07:00 localhost podman[320009]: 2025-11-27 10:07:00.007075578 +0000 UTC m=+0.107712730 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:00 localhost dnsmasq[319862]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:00 localhost podman[320010]: 2025-11-27 10:07:00.040667548 +0000 UTC m=+0.135282757 container kill 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:07:00 localhost podman[320007]: 2025-11-27 10:07:00.058014627 +0000 UTC m=+0.159960284 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, io.buildah.version=1.41.3) Nov 27 05:07:00 localhost podman[320009]: 2025-11-27 10:07:00.091032442 +0000 UTC m=+0.191669624 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, managed_by=edpm_ansible) Nov 27 05:07:00 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:07:00 localhost podman[320007]: 2025-11-27 10:07:00.142049754 +0000 UTC m=+0.243995431 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent) Nov 27 05:07:00 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:07:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:00.170 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=15, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=14) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:00 localhost nova_compute[284026]: 2025-11-27 10:07:00.170 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:00.172 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 2 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:07:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:00.443 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:00Z, description=, device_id=66e84bf5-703d-41f4-916a-e3b23da19f33, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=bcd2246f-0303-4fdd-a5e3-d419e2099966, ip_allocation=immediate, mac_address=fa:16:3e:a2:31:a9, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1532, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:00Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:00 localhost nova_compute[284026]: 2025-11-27 10:07:00.563 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:07:00 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:07:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:07:00 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:07:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:07:00 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 44ea101b-507b-44a7-b512-232e138eec05 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:07:00 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 44ea101b-507b-44a7-b512-232e138eec05 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:07:00 localhost ceph-mgr[290377]: [progress INFO root] Completed event 44ea101b-507b-44a7-b512-232e138eec05 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:07:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:07:00 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:07:00 localhost systemd[1]: tmp-crun.HfZMXE.mount: Deactivated successfully. Nov 27 05:07:01 localhost nova_compute[284026]: 2025-11-27 10:07:01.030 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:01 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:01 localhost podman[320145]: 2025-11-27 10:07:01.116397856 +0000 UTC m=+0.059608156 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:07:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v273: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:01.345 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:06:59Z, description=, device_id=b4b9bed4-45ba-4376-96cd-5363a81ae187, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=39625dd9-ac4a-41eb-a23a-3952feb589e4, ip_allocation=immediate, mac_address=fa:16:3e:15:3b:c3, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1528, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:00Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:01 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:07:01 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:07:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:01.536 265123 INFO neutron.agent.dhcp.agent [None req-5200e438-7404-4386-a543-26d6a6a3e9ef - - - - - -] DHCP configuration for ports {'bcd2246f-0303-4fdd-a5e3-d419e2099966'} is completed#033[00m Nov 27 05:07:01 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:01 localhost podman[320195]: 2025-11-27 10:07:01.689660234 +0000 UTC m=+0.076101562 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:01 localhost dnsmasq[319862]: exiting on receipt of SIGTERM Nov 27 05:07:01 localhost podman[320203]: 2025-11-27 10:07:01.718777133 +0000 UTC m=+0.071920819 container kill 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0) Nov 27 05:07:01 localhost systemd[1]: libpod-0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245.scope: Deactivated successfully. Nov 27 05:07:01 localhost podman[320228]: 2025-11-27 10:07:01.805891283 +0000 UTC m=+0.065761413 container died 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:07:01 localhost podman[320228]: 2025-11-27 10:07:01.907900595 +0000 UTC m=+0.167770675 container remove 0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:07:01 localhost systemd[1]: libpod-conmon-0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245.scope: Deactivated successfully. Nov 27 05:07:01 localhost ovn_controller[156436]: 2025-11-27T10:07:01Z|00212|binding|INFO|Releasing lport 723a44c5-5507-4969-8f33-21dec93014d9 from this chassis (sb_readonly=0) Nov 27 05:07:01 localhost kernel: device tap723a44c5-55 left promiscuous mode Nov 27 05:07:01 localhost ovn_controller[156436]: 2025-11-27T10:07:01Z|00213|binding|INFO|Setting lport 723a44c5-5507-4969-8f33-21dec93014d9 down in Southbound Nov 27 05:07:01 localhost nova_compute[284026]: 2025-11-27 10:07:01.923 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:01 localhost nova_compute[284026]: 2025-11-27 10:07:01.944 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:01 localhost systemd[1]: tmp-crun.WKyz8a.mount: Deactivated successfully. Nov 27 05:07:01 localhost systemd[1]: var-lib-containers-storage-overlay-53138c1a564f90e694ccf06cdefd8f57fbb42d0cdb23d5bde3928385505d84e2-merged.mount: Deactivated successfully. Nov 27 05:07:01 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-0aa70ca06c4fe51c8bbe4c646570232a9921df623420ac4fd0a2cef20c6b6245-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:02.007 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe5a:b6ef/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=723a44c5-5507-4969-8f33-21dec93014d9) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:02.009 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 723a44c5-5507-4969-8f33-21dec93014d9 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:02.011 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:02.012 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[722b0f70-bc81-4200-9358-cd0d05f4c92f]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:02.173 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '15'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:07:02 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:02.263 265123 INFO neutron.agent.dhcp.agent [None req-8270f31e-4a61-4782-81f5-864e7b0211a5 - - - - - -] DHCP configuration for ports {'39625dd9-ac4a-41eb-a23a-3952feb589e4'} is completed#033[00m Nov 27 05:07:02 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:02.547 265123 INFO neutron.agent.dhcp.agent [None req-c4110ea8-cd7a-4848-8e5d-c4b10dac6cc5 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:02 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v274: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:04 localhost nova_compute[284026]: 2025-11-27 10:07:04.695 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:04 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:07:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:07:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v275: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:05 localhost nova_compute[284026]: 2025-11-27 10:07:05.567 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:05 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.069 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:06 localhost podman[320277]: 2025-11-27 10:07:06.197192491 +0000 UTC m=+0.060479908 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:07:06 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:06 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:06 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:06.362 265123 INFO neutron.agent.linux.ip_lib [None req-0e548aab-cf10-464e-97cb-fee0def50e5f - - - - - -] Device tape7fc40de-89 cannot be used as it has no MAC address#033[00m Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.387 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost kernel: device tape7fc40de-89 entered promiscuous mode Nov 27 05:07:06 localhost NetworkManager[5971]: [1764238026.3964] manager: (tape7fc40de-89): new Generic device (/org/freedesktop/NetworkManager/Devices/39) Nov 27 05:07:06 localhost ovn_controller[156436]: 2025-11-27T10:07:06Z|00214|binding|INFO|Claiming lport e7fc40de-8996-43f6-a1f0-d058670c7948 for this chassis. Nov 27 05:07:06 localhost ovn_controller[156436]: 2025-11-27T10:07:06Z|00215|binding|INFO|e7fc40de-8996-43f6-a1f0-d058670c7948: Claiming unknown Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.398 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost systemd-udevd[320307]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:07:06 localhost ovn_controller[156436]: 2025-11-27T10:07:06Z|00216|binding|INFO|Setting lport e7fc40de-8996-43f6-a1f0-d058670c7948 ovn-installed in OVS Nov 27 05:07:06 localhost ovn_controller[156436]: 2025-11-27T10:07:06Z|00217|binding|INFO|Setting lport e7fc40de-8996-43f6-a1f0-d058670c7948 up in Southbound Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.407 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.409 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:06.409 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:feb6:575d/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e7fc40de-8996-43f6-a1f0-d058670c7948) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:06.411 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e7fc40de-8996-43f6-a1f0-d058670c7948 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:07:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:06.414 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port bfe82e45-ff71-4fa7-b466-01995bdd5ee4 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:06.414 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:06 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:06.415 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[7cf4ea5a-6608-458c-ae1e-97f4efc12ac1]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.443 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.483 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:06 localhost nova_compute[284026]: 2025-11-27 10:07:06.512 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v276: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:07 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:07.285 2 INFO neutron.agent.securitygroups_rpc [None req-65faea91-915b-44f3-93a9-85007b12a578 9cd13abea9f04c7c9acaacd0c9346e8c 5711c25eda684e6595773005b3d0d3af - - default default] Security group member updated ['3efaea02-6ea0-4b8a-abcb-9a7f7662a409']#033[00m Nov 27 05:07:07 localhost podman[320362]: Nov 27 05:07:07 localhost podman[320362]: 2025-11-27 10:07:07.318416733 +0000 UTC m=+0.085696093 container create 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:07:07 localhost systemd[1]: Started libpod-conmon-543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015.scope. Nov 27 05:07:07 localhost podman[320362]: 2025-11-27 10:07:07.276889518 +0000 UTC m=+0.044168908 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:07 localhost systemd[1]: tmp-crun.XZ4Z1x.mount: Deactivated successfully. Nov 27 05:07:07 localhost systemd[1]: Started libcrun container. Nov 27 05:07:07 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/db8052710150384327d2ffaec65ceb2c43f364bbcd1f4b2d2c831adcb8337802/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:07 localhost podman[320362]: 2025-11-27 10:07:07.403107977 +0000 UTC m=+0.170387347 container init 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:07:07 localhost podman[320362]: 2025-11-27 10:07:07.411294699 +0000 UTC m=+0.178574069 container start 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:07:07 localhost dnsmasq[320380]: started, version 2.85 cachesize 150 Nov 27 05:07:07 localhost dnsmasq[320380]: DNS service limited to local subnets Nov 27 05:07:07 localhost dnsmasq[320380]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:07 localhost dnsmasq[320380]: warning: no upstream servers configured Nov 27 05:07:07 localhost dnsmasq-dhcp[320380]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:07 localhost dnsmasq[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:07 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:07 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:07 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:07.542 265123 INFO neutron.agent.dhcp.agent [None req-bd152cd1-ed20-459f-a338-9c3c25f4a155 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:07 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:07.849 2 INFO neutron.agent.securitygroups_rpc [None req-cfc8d52c-5d06-426e-8522-db69cd27448d 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:07 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:07.964 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:06Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=2e2912bf-907d-492a-82d1-e24a73419140, ip_allocation=immediate, mac_address=fa:16:3e:78:d8:fc, name=tempest-NetworksTestDHCPv6-1291756885, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=40, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['2b643cf5-7531-4139-bae8-df19dacf7970'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:02Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1556, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:07Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:07:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:07:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:07:08 localhost podman[320399]: 2025-11-27 10:07:08.162931679 +0000 UTC m=+0.063877572 container kill 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:07:08 localhost dnsmasq[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:07:08 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:08 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:08 localhost podman[320412]: 2025-11-27 10:07:08.260358668 +0000 UTC m=+0.090107602 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:07:08 localhost podman[320412]: 2025-11-27 10:07:08.292866698 +0000 UTC m=+0.122615592 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:07:08 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:07:08 localhost podman[320411]: 2025-11-27 10:07:08.371073486 +0000 UTC m=+0.204706545 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible) Nov 27 05:07:08 localhost podman[320411]: 2025-11-27 10:07:08.386298459 +0000 UTC m=+0.219931518 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm) Nov 27 05:07:08 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:07:08 localhost podman[320414]: 2025-11-27 10:07:08.373852152 +0000 UTC m=+0.200853731 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, version=9.6, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.buildah.version=1.33.7, architecture=x86_64, managed_by=edpm_ansible, vcs-type=git, config_id=edpm, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9) Nov 27 05:07:08 localhost podman[320414]: 2025-11-27 10:07:08.458030992 +0000 UTC m=+0.285032551 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, architecture=x86_64, vendor=Red Hat, Inc., release=1755695350, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., managed_by=edpm_ansible, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, config_id=edpm, io.openshift.tags=minimal rhel9, vcs-type=git, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, name=ubi9-minimal, version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 05:07:08 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:07:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:08.751 265123 INFO neutron.agent.dhcp.agent [None req-474779af-6ddf-40db-9043-fcf1a772bc6e - - - - - -] DHCP configuration for ports {'2e2912bf-907d-492a-82d1-e24a73419140'} is completed#033[00m Nov 27 05:07:08 localhost podman[242678]: time="2025-11-27T10:07:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:07:08 localhost podman[242678]: @ - - [27/Nov/2025:10:07:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159329 "" "Go-http-client/1.1" Nov 27 05:07:08 localhost podman[242678]: @ - - [27/Nov/2025:10:07:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20198 "" "Go-http-client/1.1" Nov 27 05:07:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:09.017 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:08Z, description=, device_id=d75c8bc0-b87a-434c-9504-55c059d1f024, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=ed1811c9-0e91-4b99-8d73-882f753fd39a, ip_allocation=immediate, mac_address=fa:16:3e:57:4e:b6, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1565, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:08Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v277: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:09 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:09 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:09 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:09 localhost podman[320496]: 2025-11-27 10:07:09.226110408 +0000 UTC m=+0.060636064 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:07:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:09.386 265123 INFO neutron.agent.dhcp.agent [None req-62d2ec59-9ee5-412d-b7f9-e28be7c585ec - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:07Z, description=, device_id=959f6973-d259-4191-984f-4d67307a3317, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=a3b7b4b7-6a3c-4043-9739-38c0049b8866, ip_allocation=immediate, mac_address=fa:16:3e:73:12:ff, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1563, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:08Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:09.546 265123 INFO neutron.agent.dhcp.agent [None req-31237508-fa2e-494d-9116-f7e328764b49 - - - - - -] DHCP configuration for ports {'ed1811c9-0e91-4b99-8d73-882f753fd39a'} is completed#033[00m Nov 27 05:07:09 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:07:09 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:09 localhost podman[320532]: 2025-11-27 10:07:09.609947151 +0000 UTC m=+0.062532625 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:07:09 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:10 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:10.025 265123 INFO neutron.agent.dhcp.agent [None req-592813f9-01da-49a4-b9d5-a43a9db7e7f8 - - - - - -] DHCP configuration for ports {'a3b7b4b7-6a3c-4043-9739-38c0049b8866'} is completed#033[00m Nov 27 05:07:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:10.158 2 INFO neutron.agent.securitygroups_rpc [None req-ee1b2c43-df85-4b5f-89a9-2ab54aceb361 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:10 localhost podman[320571]: 2025-11-27 10:07:10.399706187 +0000 UTC m=+0.057877470 container kill 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:10 localhost dnsmasq[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:10 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:10 localhost dnsmasq-dhcp[320380]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:10 localhost nova_compute[284026]: 2025-11-27 10:07:10.570 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:11 localhost nova_compute[284026]: 2025-11-27 10:07:11.112 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v278: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:11 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:11 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:11 localhost podman[320611]: 2025-11-27 10:07:11.60315815 +0000 UTC m=+0.070175300 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:07:11 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:11 localhost systemd[1]: tmp-crun.ESrlzS.mount: Deactivated successfully. Nov 27 05:07:12 localhost dnsmasq[320380]: exiting on receipt of SIGTERM Nov 27 05:07:12 localhost podman[320650]: 2025-11-27 10:07:12.108810976 +0000 UTC m=+0.063671436 container kill 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:07:12 localhost systemd[1]: libpod-543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015.scope: Deactivated successfully. Nov 27 05:07:12 localhost podman[320662]: 2025-11-27 10:07:12.183275211 +0000 UTC m=+0.062411692 container died 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:12 localhost podman[320662]: 2025-11-27 10:07:12.214851841 +0000 UTC m=+0.093988272 container cleanup 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:07:12 localhost systemd[1]: libpod-conmon-543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015.scope: Deactivated successfully. Nov 27 05:07:12 localhost podman[320664]: 2025-11-27 10:07:12.26792266 +0000 UTC m=+0.137606946 container remove 543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:12 localhost nova_compute[284026]: 2025-11-27 10:07:12.283 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:12 localhost ovn_controller[156436]: 2025-11-27T10:07:12Z|00218|binding|INFO|Releasing lport e7fc40de-8996-43f6-a1f0-d058670c7948 from this chassis (sb_readonly=0) Nov 27 05:07:12 localhost ovn_controller[156436]: 2025-11-27T10:07:12Z|00219|binding|INFO|Setting lport e7fc40de-8996-43f6-a1f0-d058670c7948 down in Southbound Nov 27 05:07:12 localhost kernel: device tape7fc40de-89 left promiscuous mode Nov 27 05:07:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:12.292 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:feb6:575d/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e7fc40de-8996-43f6-a1f0-d058670c7948) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:12.294 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e7fc40de-8996-43f6-a1f0-d058670c7948 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:12.296 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:12.297 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3a543250-1cc1-4c7a-a645-3876da5c2504]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:12 localhost nova_compute[284026]: 2025-11-27 10:07:12.311 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:12 localhost systemd[1]: var-lib-containers-storage-overlay-db8052710150384327d2ffaec65ceb2c43f364bbcd1f4b2d2c831adcb8337802-merged.mount: Deactivated successfully. Nov 27 05:07:12 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-543cfe078bdecaecf50ac477ab509fb13763eeb6e73370b086acbaecc3b33015-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:12 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:12 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:12.640 265123 INFO neutron.agent.dhcp.agent [None req-ea3efd24-e61a-489c-b6af-fad865d67a91 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v279: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:14.106 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:13Z, description=, device_id=9913af50-31f6-4b91-9426-b698e666ebea, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=2357a353-cfbd-4e65-98d6-2eec09587e30, ip_allocation=immediate, mac_address=fa:16:3e:f3:04:8b, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1587, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:13Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:14 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:07:14 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:14 localhost podman[320711]: 2025-11-27 10:07:14.39263059 +0000 UTC m=+0.065516996 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:07:14 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:14.564 265123 INFO neutron.agent.linux.ip_lib [None req-59764f5f-c9b0-4dd2-9577-a04be312d2f4 - - - - - -] Device tapa1ed1748-6b cannot be used as it has no MAC address#033[00m Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.641 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:14 localhost kernel: device tapa1ed1748-6b entered promiscuous mode Nov 27 05:07:14 localhost NetworkManager[5971]: [1764238034.6505] manager: (tapa1ed1748-6b): new Generic device (/org/freedesktop/NetworkManager/Devices/40) Nov 27 05:07:14 localhost ovn_controller[156436]: 2025-11-27T10:07:14Z|00220|binding|INFO|Claiming lport a1ed1748-6bd1-4b84-a151-e4d25718fe18 for this chassis. Nov 27 05:07:14 localhost ovn_controller[156436]: 2025-11-27T10:07:14Z|00221|binding|INFO|a1ed1748-6bd1-4b84-a151-e4d25718fe18: Claiming unknown Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.655 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:14 localhost systemd-udevd[320743]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:07:14 localhost ovn_controller[156436]: 2025-11-27T10:07:14Z|00222|binding|INFO|Setting lport a1ed1748-6bd1-4b84-a151-e4d25718fe18 ovn-installed in OVS Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.665 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:14 localhost ovn_controller[156436]: 2025-11-27T10:07:14Z|00223|binding|INFO|Setting lport a1ed1748-6bd1-4b84-a151-e4d25718fe18 up in Southbound Nov 27 05:07:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:14.670 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe69:8e8d/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=a1ed1748-6bd1-4b84-a151-e4d25718fe18) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:14.672 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a1ed1748-6bd1-4b84-a151-e4d25718fe18 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:07:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:14.674 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port a54d0e6e-a259-4382-a9ec-cb0f2eee1676 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:14.674 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:14.675 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c37a380b-be2e-46d2-bc74-51102354bc4c]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.687 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:14.693 265123 INFO neutron.agent.dhcp.agent [None req-ca8f6c70-3d2c-4ecc-8b9c-75d933ef7f4a - - - - - -] DHCP configuration for ports {'2357a353-cfbd-4e65-98d6-2eec09587e30'} is completed#033[00m Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost journal[232028]: ethtool ioctl error on tapa1ed1748-6b: No such device Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.732 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:14 localhost nova_compute[284026]: 2025-11-27 10:07:14.765 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:15.020 2 INFO neutron.agent.securitygroups_rpc [None req-f6b79fdf-2578-4798-b54c-4b3e64bad548 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v280: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:15 localhost nova_compute[284026]: 2025-11-27 10:07:15.573 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:15 localhost podman[320814]: Nov 27 05:07:15 localhost podman[320814]: 2025-11-27 10:07:15.628564439 +0000 UTC m=+0.082866983 container create c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:07:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:07:15 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:07:15 localhost systemd[1]: Started libpod-conmon-c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d.scope. Nov 27 05:07:15 localhost podman[320814]: 2025-11-27 10:07:15.593563517 +0000 UTC m=+0.047866101 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:15 localhost systemd[1]: Started libcrun container. Nov 27 05:07:15 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/02bb3882cf243b96523bf88c50f0c0460a6f2581e5cce76435dbce5ac422d411/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:15 localhost podman[320814]: 2025-11-27 10:07:15.721057809 +0000 UTC m=+0.175360363 container init c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:15 localhost dnsmasq[320855]: started, version 2.85 cachesize 150 Nov 27 05:07:15 localhost dnsmasq[320855]: DNS service limited to local subnets Nov 27 05:07:15 localhost dnsmasq[320855]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:15 localhost dnsmasq[320855]: warning: no upstream servers configured Nov 27 05:07:15 localhost dnsmasq[320855]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:15 localhost podman[320814]: 2025-11-27 10:07:15.746439013 +0000 UTC m=+0.200741537 container start c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:15 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:15.793 265123 INFO neutron.agent.dhcp.agent [None req-59764f5f-c9b0-4dd2-9577-a04be312d2f4 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:14Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=421655b4-49eb-4671-b89a-59b5789f3f32, ip_allocation=immediate, mac_address=fa:16:3e:a1:e7:59, name=tempest-NetworksTestDHCPv6-104432580, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=42, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['f9fd1f90-f75f-48b8-bfc2-d5324675630d'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:12Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1589, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:14Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:15 localhost podman[320829]: 2025-11-27 10:07:15.820816576 +0000 UTC m=+0.141572774 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:07:15 localhost podman[320829]: 2025-11-27 10:07:15.83285738 +0000 UTC m=+0.153613568 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:07:15 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:07:15 localhost podman[320828]: 2025-11-27 10:07:15.87521149 +0000 UTC m=+0.198733842 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:07:15 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:15.901 265123 INFO neutron.agent.dhcp.agent [None req-2474d9ed-a6c0-4cb4-a9dc-83a584138bd9 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:15 localhost podman[320828]: 2025-11-27 10:07:15.994063081 +0000 UTC m=+0.317585503 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 05:07:16 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:07:16 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:16 localhost podman[320925]: 2025-11-27 10:07:16.030645136 +0000 UTC m=+0.057713786 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:16 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:16 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:16 localhost dnsmasq[320855]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:07:16 localhost podman[320911]: 2025-11-27 10:07:16.054883758 +0000 UTC m=+0.109593582 container kill c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:16 localhost nova_compute[284026]: 2025-11-27 10:07:16.115 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:16.287 2 INFO neutron.agent.securitygroups_rpc [None req-b9955938-6bd1-4511-af28-9ff091a51dab 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:16 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:16.321 265123 INFO neutron.agent.dhcp.agent [None req-a1b52fe3-436a-4c00-be59-c115ce479f0e - - - - - -] DHCP configuration for ports {'421655b4-49eb-4671-b89a-59b5789f3f32'} is completed#033[00m Nov 27 05:07:16 localhost systemd[1]: tmp-crun.MVvPUM.mount: Deactivated successfully. Nov 27 05:07:16 localhost nova_compute[284026]: 2025-11-27 10:07:16.781 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:16 localhost dnsmasq[320855]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:16 localhost podman[320977]: 2025-11-27 10:07:16.867406356 +0000 UTC m=+0.048665601 container kill c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:07:17 localhost nova_compute[284026]: 2025-11-27 10:07:17.011 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v281: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:17 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:17 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:17 localhost podman[321015]: 2025-11-27 10:07:17.590621389 +0000 UTC m=+0.066129151 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:07:17 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:17 localhost dnsmasq[320855]: exiting on receipt of SIGTERM Nov 27 05:07:17 localhost podman[321046]: 2025-11-27 10:07:17.74699169 +0000 UTC m=+0.073664054 container kill c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:17 localhost systemd[1]: libpod-c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d.scope: Deactivated successfully. Nov 27 05:07:17 localhost ovn_controller[156436]: 2025-11-27T10:07:17Z|00224|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:07:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:17 localhost podman[321064]: 2025-11-27 10:07:17.822354369 +0000 UTC m=+0.057888540 container died c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:07:17 localhost nova_compute[284026]: 2025-11-27 10:07:17.841 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:17 localhost podman[321064]: 2025-11-27 10:07:17.860189379 +0000 UTC m=+0.095723510 container cleanup c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:17 localhost systemd[1]: libpod-conmon-c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d.scope: Deactivated successfully. Nov 27 05:07:17 localhost podman[321066]: 2025-11-27 10:07:17.905974941 +0000 UTC m=+0.135154260 container remove c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:07:17 localhost ovn_controller[156436]: 2025-11-27T10:07:17Z|00225|binding|INFO|Releasing lport a1ed1748-6bd1-4b84-a151-e4d25718fe18 from this chassis (sb_readonly=0) Nov 27 05:07:17 localhost ovn_controller[156436]: 2025-11-27T10:07:17Z|00226|binding|INFO|Setting lport a1ed1748-6bd1-4b84-a151-e4d25718fe18 down in Southbound Nov 27 05:07:17 localhost kernel: device tapa1ed1748-6b left promiscuous mode Nov 27 05:07:17 localhost nova_compute[284026]: 2025-11-27 10:07:17.918 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:17 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:17.930 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe69:8e8d/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=a1ed1748-6bd1-4b84-a151-e4d25718fe18) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:17 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:17.932 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a1ed1748-6bd1-4b84-a151-e4d25718fe18 in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:17 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:17.936 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:17 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:17.937 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4a4fb039-197c-4c99-86d6-98594bb7a5e9]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:17 localhost nova_compute[284026]: 2025-11-27 10:07:17.946 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:18.258 265123 INFO neutron.agent.dhcp.agent [None req-3fa13f75-40f4-49f7-ad90-676cecb607f2 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:18.529 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:18Z, description=, device_id=6e94f718-d486-4336-a41a-65ab798868e4, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=4c315f7f-2c63-431f-944b-a13d5446ca59, ip_allocation=immediate, mac_address=fa:16:3e:15:3f:fb, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1595, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:18Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:18 localhost nova_compute[284026]: 2025-11-27 10:07:18.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:18 localhost systemd[1]: tmp-crun.BsqItP.mount: Deactivated successfully. Nov 27 05:07:18 localhost systemd[1]: var-lib-containers-storage-overlay-02bb3882cf243b96523bf88c50f0c0460a6f2581e5cce76435dbce5ac422d411-merged.mount: Deactivated successfully. Nov 27 05:07:18 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c80a9fd8d426c7f4eb1f339b82f95a108fbc1d8bdc97c915dde2effd9b6d5e1d-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:18 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:18 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:18 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:18 localhost podman[321108]: 2025-11-27 10:07:18.754123658 +0000 UTC m=+0.060727556 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:18 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:19 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:19.157 265123 INFO neutron.agent.dhcp.agent [None req-d167356a-28d6-482e-9cb5-23288e4acb0b - - - - - -] DHCP configuration for ports {'4c315f7f-2c63-431f-944b-a13d5446ca59'} is completed#033[00m Nov 27 05:07:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v282: 177 pgs: 177 active+clean; 145 MiB data, 842 MiB used, 41 GiB / 42 GiB avail Nov 27 05:07:19 localhost nova_compute[284026]: 2025-11-27 10:07:19.691 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:20.521 265123 INFO neutron.agent.linux.ip_lib [None req-170f2993-c701-46d8-b5c4-948f61dc0854 - - - - - -] Device tap47d8782a-e7 cannot be used as it has no MAC address#033[00m Nov 27 05:07:20 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:20.528 2 INFO neutron.agent.securitygroups_rpc [None req-f5f8230f-c132-46cf-b142-5837d7358f7a 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.547 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost kernel: device tap47d8782a-e7 entered promiscuous mode Nov 27 05:07:20 localhost NetworkManager[5971]: [1764238040.5549] manager: (tap47d8782a-e7): new Generic device (/org/freedesktop/NetworkManager/Devices/41) Nov 27 05:07:20 localhost ovn_controller[156436]: 2025-11-27T10:07:20Z|00227|binding|INFO|Claiming lport 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd for this chassis. Nov 27 05:07:20 localhost ovn_controller[156436]: 2025-11-27T10:07:20Z|00228|binding|INFO|47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd: Claiming unknown Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.555 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost systemd-udevd[321140]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:07:20 localhost ovn_controller[156436]: 2025-11-27T10:07:20Z|00229|binding|INFO|Setting lport 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd ovn-installed in OVS Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.564 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.568 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost ovn_controller[156436]: 2025-11-27T10:07:20Z|00230|binding|INFO|Setting lport 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd up in Southbound Nov 27 05:07:20 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:20.570 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:febd:e126/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:20 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:20.571 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:07:20 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:20.572 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 97190f8e-1daa-4516-8737-635d7febf9b5 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:20 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:20.573 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.574 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:20.574 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0cd01e67-e4ea-498b-bac2-f22d8c9c8d1b]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.591 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.629 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:20 localhost nova_compute[284026]: 2025-11-27 10:07:20.660 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:21 localhost nova_compute[284026]: 2025-11-27 10:07:21.159 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v283: 177 pgs: 177 active+clean; 145 MiB data, 834 MiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 341 B/s wr, 0 op/s Nov 27 05:07:21 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:21 localhost podman[321189]: 2025-11-27 10:07:21.212832661 +0000 UTC m=+0.080436796 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:21 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:21 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:21 localhost ovn_controller[156436]: 2025-11-27T10:07:21Z|00231|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:07:21 localhost nova_compute[284026]: 2025-11-27 10:07:21.275 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:21 localhost podman[321235]: Nov 27 05:07:21 localhost podman[321235]: 2025-11-27 10:07:21.632318706 +0000 UTC m=+0.106290883 container create 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:21 localhost podman[321235]: 2025-11-27 10:07:21.578386334 +0000 UTC m=+0.052358551 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:21 localhost systemd[1]: Started libpod-conmon-21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6.scope. Nov 27 05:07:21 localhost systemd[1]: Started libcrun container. Nov 27 05:07:21 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/1d188bce9db2d1ec98629b7e364bc47f276002a67e0c1c826fa0e6d1f6c96855/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:21 localhost podman[321235]: 2025-11-27 10:07:21.714710575 +0000 UTC m=+0.188682762 container init 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:07:21 localhost podman[321235]: 2025-11-27 10:07:21.725400172 +0000 UTC m=+0.199372349 container start 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:07:21 localhost nova_compute[284026]: 2025-11-27 10:07:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:21 localhost dnsmasq[321254]: started, version 2.85 cachesize 150 Nov 27 05:07:21 localhost dnsmasq[321254]: DNS service limited to local subnets Nov 27 05:07:21 localhost dnsmasq[321254]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:21 localhost dnsmasq[321254]: warning: no upstream servers configured Nov 27 05:07:21 localhost dnsmasq-dhcp[321254]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:21 localhost dnsmasq[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:21 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:21 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:21 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:21.796 265123 INFO neutron.agent.dhcp.agent [None req-170f2993-c701-46d8-b5c4-948f61dc0854 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:20Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=f0da7762-b3bf-4330-8d0f-8011c2c16045, ip_allocation=immediate, mac_address=fa:16:3e:7c:1e:ca, name=tempest-NetworksTestDHCPv6-1445115304, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=44, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['ededd636-ed3d-4562-9b99-04265c27825d'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:17Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1597, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:20Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:22 localhost dnsmasq[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 1 addresses Nov 27 05:07:22 localhost podman[321273]: 2025-11-27 10:07:22.007491978 +0000 UTC m=+0.071830415 container kill 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:07:22 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:22 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:22.092 265123 INFO neutron.agent.dhcp.agent [None req-b9c7b8d4-61db-4f98-bafd-d79b9f3f4a15 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:22 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:22.107 2 INFO neutron.agent.securitygroups_rpc [None req-73ea3034-fd9e-4d8a-8aff-0b9e610c1d87 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:22 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:22.293 265123 INFO neutron.agent.dhcp.agent [None req-413eec15-17d6-4a75-af56-0f444d2587dc - - - - - -] DHCP configuration for ports {'f0da7762-b3bf-4330-8d0f-8011c2c16045'} is completed#033[00m Nov 27 05:07:22 localhost dnsmasq[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:22 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:22 localhost dnsmasq-dhcp[321254]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:22 localhost podman[321310]: 2025-11-27 10:07:22.38084407 +0000 UTC m=+0.065835863 container kill 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:07:22 localhost nova_compute[284026]: 2025-11-27 10:07:22.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:22 localhost ovn_controller[156436]: 2025-11-27T10:07:22Z|00232|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:07:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e140 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:22 localhost nova_compute[284026]: 2025-11-27 10:07:22.863 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:22 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:07:22 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:22 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:22 localhost podman[321347]: 2025-11-27 10:07:22.872690184 +0000 UTC m=+0.193525151 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:23 localhost dnsmasq[321254]: exiting on receipt of SIGTERM Nov 27 05:07:23 localhost podman[321383]: 2025-11-27 10:07:23.024620475 +0000 UTC m=+0.055770393 container kill 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:23 localhost systemd[1]: libpod-21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6.scope: Deactivated successfully. Nov 27 05:07:23 localhost podman[321406]: 2025-11-27 10:07:23.092215895 +0000 UTC m=+0.047107170 container died 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:07:23 localhost podman[321406]: 2025-11-27 10:07:23.189461004 +0000 UTC m=+0.144352249 container remove 21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:07:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v284: 177 pgs: 177 active+clean; 145 MiB data, 834 MiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 341 B/s wr, 0 op/s Nov 27 05:07:23 localhost systemd[1]: libpod-conmon-21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6.scope: Deactivated successfully. Nov 27 05:07:23 localhost nova_compute[284026]: 2025-11-27 10:07:23.208 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:23 localhost ovn_controller[156436]: 2025-11-27T10:07:23Z|00233|binding|INFO|Releasing lport 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd from this chassis (sb_readonly=0) Nov 27 05:07:23 localhost ovn_controller[156436]: 2025-11-27T10:07:23Z|00234|binding|INFO|Setting lport 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd down in Southbound Nov 27 05:07:23 localhost kernel: device tap47d8782a-e7 left promiscuous mode Nov 27 05:07:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:23.220 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:febd:e126/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:23.221 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 47d8782a-e7d5-4e6b-a0c1-8a030f70b0bd in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:23.222 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:23.223 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a9162102-9d23-4cf2-8533-20db79fbc502]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:23 localhost nova_compute[284026]: 2025-11-27 10:07:23.236 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:23.340 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:23Z, description=, device_id=79a5dcac-0d92-4f52-8bab-f9555eaf8ff1, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=28a83e87-0e81-40a8-9c17-6cbaf54b2f1d, ip_allocation=immediate, mac_address=fa:16:3e:e1:c1:ca, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1604, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:23Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:23 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:23 localhost podman[321443]: 2025-11-27 10:07:23.606499963 +0000 UTC m=+0.058481996 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:23 localhost systemd[1]: var-lib-containers-storage-overlay-1d188bce9db2d1ec98629b7e364bc47f276002a67e0c1c826fa0e6d1f6c96855-merged.mount: Deactivated successfully. Nov 27 05:07:23 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-21d7fcc8cde93da1374f0f136d111284a795e27d739ea22316ec6bbd7d95cac6-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:23 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:23 localhost nova_compute[284026]: 2025-11-27 10:07:23.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:23 localhost nova_compute[284026]: 2025-11-27 10:07:23.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:23 localhost nova_compute[284026]: 2025-11-27 10:07:23.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:07:23 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:23.741 2 INFO neutron.agent.securitygroups_rpc [None req-9ee64373-fa7c-4a83-b0c3-0eb990f647c4 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e141 e141: 6 total, 6 up, 6 in Nov 27 05:07:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:23.895 265123 INFO neutron.agent.dhcp.agent [None req-5d04883a-513a-4943-b3fd-6b6245f05462 - - - - - -] DHCP configuration for ports {'28a83e87-0e81-40a8-9c17-6cbaf54b2f1d'} is completed#033[00m Nov 27 05:07:24 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:24.247 2 INFO neutron.agent.securitygroups_rpc [None req-27fa51dc-eba9-4b80-a4eb-2b587db31305 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:24.350 265123 INFO neutron.agent.linux.ip_lib [None req-feeb53d6-b897-4eca-96a9-59dcfb2420a0 - - - - - -] Device tap79f55ca3-e6 cannot be used as it has no MAC address#033[00m Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.372 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost kernel: device tap79f55ca3-e6 entered promiscuous mode Nov 27 05:07:24 localhost NetworkManager[5971]: [1764238044.3809] manager: (tap79f55ca3-e6): new Generic device (/org/freedesktop/NetworkManager/Devices/42) Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.380 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost ovn_controller[156436]: 2025-11-27T10:07:24Z|00235|binding|INFO|Claiming lport 79f55ca3-e6a0-4338-902b-abfec6942f3d for this chassis. Nov 27 05:07:24 localhost ovn_controller[156436]: 2025-11-27T10:07:24Z|00236|binding|INFO|79f55ca3-e6a0-4338-902b-abfec6942f3d: Claiming unknown Nov 27 05:07:24 localhost systemd-udevd[321475]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:07:24 localhost ovn_controller[156436]: 2025-11-27T10:07:24Z|00237|binding|INFO|Setting lport 79f55ca3-e6a0-4338-902b-abfec6942f3d ovn-installed in OVS Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.390 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.394 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:24.394 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=79f55ca3-e6a0-4338-902b-abfec6942f3d) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:24.396 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 79f55ca3-e6a0-4338-902b-abfec6942f3d in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:07:24 localhost ovn_controller[156436]: 2025-11-27T10:07:24Z|00238|binding|INFO|Setting lport 79f55ca3-e6a0-4338-902b-abfec6942f3d up in Southbound Nov 27 05:07:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:24.398 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 4800a92e-9b20-4b82-ad6f-d743029df286 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:24.398 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:24.399 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[f19d1e79-f9c6-4692-bfe0-9b961e435b4f]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.415 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost journal[232028]: ethtool ioctl error on tap79f55ca3-e6: No such device Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.461 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:24 localhost nova_compute[284026]: 2025-11-27 10:07:24.491 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v286: 177 pgs: 177 active+clean; 145 MiB data, 834 MiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 409 B/s wr, 1 op/s Nov 27 05:07:25 localhost podman[321545]: Nov 27 05:07:25 localhost podman[321545]: 2025-11-27 10:07:25.336612168 +0000 UTC m=+0.096537650 container create d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:07:25 localhost systemd[1]: Started libpod-conmon-d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60.scope. Nov 27 05:07:25 localhost podman[321545]: 2025-11-27 10:07:25.29247005 +0000 UTC m=+0.052395522 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:25 localhost systemd[1]: Started libcrun container. Nov 27 05:07:25 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/9bf42b1fa96ae4f95b03a5cddecac1d69f7bdc8dbb039a89fa8cfd7e106b6753/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:25 localhost podman[321545]: 2025-11-27 10:07:25.423617441 +0000 UTC m=+0.183542903 container init d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:07:25 localhost dnsmasq[321563]: started, version 2.85 cachesize 150 Nov 27 05:07:25 localhost dnsmasq[321563]: DNS service limited to local subnets Nov 27 05:07:25 localhost dnsmasq[321563]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:25 localhost dnsmasq[321563]: warning: no upstream servers configured Nov 27 05:07:25 localhost dnsmasq-dhcp[321563]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:25 localhost podman[321545]: 2025-11-27 10:07:25.456845645 +0000 UTC m=+0.216771167 container start d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:07:25 localhost dnsmasq[321563]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:25 localhost dnsmasq-dhcp[321563]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:25 localhost dnsmasq-dhcp[321563]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:25 localhost nova_compute[284026]: 2025-11-27 10:07:25.578 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:25 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:25.608 265123 INFO neutron.agent.dhcp.agent [None req-a221968f-a7f8-4af7-9d86-504a4d4f65a2 - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:25 localhost nova_compute[284026]: 2025-11-27 10:07:25.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:25 localhost openstack_network_exporter[244641]: ERROR 10:07:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:07:25 localhost openstack_network_exporter[244641]: ERROR 10:07:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:07:25 localhost openstack_network_exporter[244641]: ERROR 10:07:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:07:25 localhost openstack_network_exporter[244641]: Nov 27 05:07:25 localhost openstack_network_exporter[244641]: ERROR 10:07:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:07:25 localhost openstack_network_exporter[244641]: ERROR 10:07:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:07:25 localhost openstack_network_exporter[244641]: Nov 27 05:07:25 localhost podman[321581]: 2025-11-27 10:07:25.81020527 +0000 UTC m=+0.063911802 container kill d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:25 localhost dnsmasq[321563]: exiting on receipt of SIGTERM Nov 27 05:07:25 localhost systemd[1]: libpod-d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60.scope: Deactivated successfully. Nov 27 05:07:25 localhost podman[321596]: 2025-11-27 10:07:25.886479914 +0000 UTC m=+0.057843129 container died d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:07:25 localhost podman[321596]: 2025-11-27 10:07:25.912931246 +0000 UTC m=+0.084294421 container cleanup d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:07:25 localhost systemd[1]: libpod-conmon-d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60.scope: Deactivated successfully. Nov 27 05:07:26 localhost podman[321597]: 2025-11-27 10:07:26.006698431 +0000 UTC m=+0.173016320 container remove d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:26 localhost nova_compute[284026]: 2025-11-27 10:07:26.066 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:26 localhost kernel: device tap79f55ca3-e6 left promiscuous mode Nov 27 05:07:26 localhost ovn_controller[156436]: 2025-11-27T10:07:26Z|00239|binding|INFO|Releasing lport 79f55ca3-e6a0-4338-902b-abfec6942f3d from this chassis (sb_readonly=0) Nov 27 05:07:26 localhost ovn_controller[156436]: 2025-11-27T10:07:26Z|00240|binding|INFO|Setting lport 79f55ca3-e6a0-4338-902b-abfec6942f3d down in Southbound Nov 27 05:07:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:26.078 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '4', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=3, gateway_chassis=[], requested_chassis=[], logical_port=79f55ca3-e6a0-4338-902b-abfec6942f3d) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:26.080 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 79f55ca3-e6a0-4338-902b-abfec6942f3d in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:26.083 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:26.084 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8e15467d-9982-4f5e-8dc3-5e0fccd13730]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:26 localhost nova_compute[284026]: 2025-11-27 10:07:26.093 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:26 localhost nova_compute[284026]: 2025-11-27 10:07:26.163 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:26 localhost systemd[1]: var-lib-containers-storage-overlay-9bf42b1fa96ae4f95b03a5cddecac1d69f7bdc8dbb039a89fa8cfd7e106b6753-merged.mount: Deactivated successfully. Nov 27 05:07:26 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-d44ed40b5606676df5f3a2f95bbb89ed891a27ac8c425d8477173d4342dc9a60-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:26.543 265123 INFO neutron.agent.dhcp.agent [None req-e6ba3c06-dbf1-4fc5-8e29-d8f0c6a20464 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:26.544 265123 INFO neutron.agent.dhcp.agent [None req-e6ba3c06-dbf1-4fc5-8e29-d8f0c6a20464 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:26.544 265123 INFO neutron.agent.dhcp.agent [None req-e6ba3c06-dbf1-4fc5-8e29-d8f0c6a20464 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:26 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:26 localhost nova_compute[284026]: 2025-11-27 10:07:26.726 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v287: 177 pgs: 177 active+clean; 201 MiB data, 930 MiB used, 41 GiB / 42 GiB avail; 35 KiB/s rd, 5.6 MiB/s wr, 49 op/s Nov 27 05:07:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e141 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:28 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:28.018 2 INFO neutron.agent.securitygroups_rpc [None req-96c93239-1833-49bc-9b57-ea419f275a3e f3e59f86157c4a87bcd419c08b2737a3 199c995d53164d408ada510b321b5483 - - default default] Security group member updated ['2244145e-f58a-48f3-bccf-8e97f62f0a77']#033[00m Nov 27 05:07:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:28.087 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:27Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=14aad6e3-66d4-4811-98b1-9e02bb490c82, ip_allocation=immediate, mac_address=fa:16:3e:c4:38:a7, name=tempest-RoutersAdminNegativeIpV6Test-1730028686, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=True, project_id=199c995d53164d408ada510b321b5483, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['2244145e-f58a-48f3-bccf-8e97f62f0a77'], standard_attr_id=1624, status=DOWN, tags=[], tenant_id=199c995d53164d408ada510b321b5483, updated_at=2025-11-27T10:07:27Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:28 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4182362587' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:28 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4182362587' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:28 localhost podman[321642]: 2025-11-27 10:07:28.365392781 +0000 UTC m=+0.078746201 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 05:07:28 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:28 localhost nova_compute[284026]: 2025-11-27 10:07:28.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:28.757 265123 INFO neutron.agent.dhcp.agent [None req-8b33523e-e446-4c8a-85c3-a8e81172ac2e - - - - - -] DHCP configuration for ports {'14aad6e3-66d4-4811-98b1-9e02bb490c82'} is completed#033[00m Nov 27 05:07:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:28.956 265123 INFO neutron.agent.linux.ip_lib [None req-e8cd6c58-0e65-40ed-8170-a2adcb97c1a6 - - - - - -] Device tap03eab131-d1 cannot be used as it has no MAC address#033[00m Nov 27 05:07:28 localhost nova_compute[284026]: 2025-11-27 10:07:28.978 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:28 localhost kernel: device tap03eab131-d1 entered promiscuous mode Nov 27 05:07:28 localhost NetworkManager[5971]: [1764238048.9848] manager: (tap03eab131-d1): new Generic device (/org/freedesktop/NetworkManager/Devices/43) Nov 27 05:07:28 localhost ovn_controller[156436]: 2025-11-27T10:07:28Z|00241|binding|INFO|Claiming lport 03eab131-d152-4149-a437-e20cd160cc9e for this chassis. Nov 27 05:07:28 localhost ovn_controller[156436]: 2025-11-27T10:07:28Z|00242|binding|INFO|03eab131-d152-4149-a437-e20cd160cc9e: Claiming unknown Nov 27 05:07:28 localhost nova_compute[284026]: 2025-11-27 10:07:28.985 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:28 localhost systemd-udevd[321673]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:07:28 localhost ovn_controller[156436]: 2025-11-27T10:07:28Z|00243|binding|INFO|Setting lport 03eab131-d152-4149-a437-e20cd160cc9e ovn-installed in OVS Nov 27 05:07:29 localhost nova_compute[284026]: 2025-11-27 10:07:28.998 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:29 localhost ovn_controller[156436]: 2025-11-27T10:07:29Z|00244|binding|INFO|Setting lport 03eab131-d152-4149-a437-e20cd160cc9e up in Southbound Nov 27 05:07:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:29.007 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fea5:4138/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=03eab131-d152-4149-a437-e20cd160cc9e) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:29 localhost nova_compute[284026]: 2025-11-27 10:07:29.007 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:29.010 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 03eab131-d152-4149-a437-e20cd160cc9e in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd bound to our chassis#033[00m Nov 27 05:07:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:29.013 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port ed52eec8-9f5d-45db-b72d-a5c9e33828a3 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:29.013 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:29.014 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[13ebbdff-986f-4d29-a329-a76c726e9b55]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:29 localhost nova_compute[284026]: 2025-11-27 10:07:29.035 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:29 localhost nova_compute[284026]: 2025-11-27 10:07:29.075 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:29 localhost nova_compute[284026]: 2025-11-27 10:07:29.106 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v288: 177 pgs: 177 active+clean; 201 MiB data, 930 MiB used, 41 GiB / 42 GiB avail; 35 KiB/s rd, 5.6 MiB/s wr, 49 op/s Nov 27 05:07:29 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:29 localhost podman[321706]: 2025-11-27 10:07:29.342741738 +0000 UTC m=+0.067427278 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:29 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:29 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:30 localhost podman[321766]: Nov 27 05:07:30 localhost podman[321766]: 2025-11-27 10:07:30.103728667 +0000 UTC m=+0.088974527 container create 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:07:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:07:30 localhost systemd[1]: Started libpod-conmon-28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae.scope. Nov 27 05:07:30 localhost podman[321766]: 2025-11-27 10:07:30.054320518 +0000 UTC m=+0.039566378 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:30 localhost systemd[1]: Started libcrun container. Nov 27 05:07:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:07:30 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4278585804dce584c080adf6f09d571c2f2a079678408ec63f158a50867a242f/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:30 localhost podman[321766]: 2025-11-27 10:07:30.180365461 +0000 UTC m=+0.165611321 container init 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:07:30 localhost podman[321766]: 2025-11-27 10:07:30.190652958 +0000 UTC m=+0.175898818 container start 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:07:30 localhost dnsmasq[321799]: started, version 2.85 cachesize 150 Nov 27 05:07:30 localhost dnsmasq[321799]: DNS service limited to local subnets Nov 27 05:07:30 localhost dnsmasq[321799]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:30 localhost dnsmasq[321799]: warning: no upstream servers configured Nov 27 05:07:30 localhost dnsmasq[321799]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:30 localhost podman[321780]: 2025-11-27 10:07:30.248929727 +0000 UTC m=+0.100287081 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, config_id=multipathd, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, container_name=multipathd, org.label-schema.vendor=CentOS) Nov 27 05:07:30 localhost podman[321780]: 2025-11-27 10:07:30.304126784 +0000 UTC m=+0.155484148 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible) Nov 27 05:07:30 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:07:30 localhost podman[321792]: 2025-11-27 10:07:30.32184531 +0000 UTC m=+0.142419816 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:07:30 localhost podman[321792]: 2025-11-27 10:07:30.359010391 +0000 UTC m=+0.179584877 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:07:30 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:07:30 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:30.370 265123 INFO neutron.agent.dhcp.agent [None req-dcbaa3d2-fc96-4e5c-a913-0f71852ab95b - - - - - -] DHCP configuration for ports {'2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:30 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:30.436 2 INFO neutron.agent.securitygroups_rpc [None req-addee7b1-1ec5-48a3-88c2-40b125cfc5d5 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:30 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:30.465 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:29Z, description=, device_id=1c9665b1-40b0-4dfb-a25d-5f9ec828cb1a, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=eaed8bc9-3c9b-4b50-a469-54f09e0b4dcc, ip_allocation=immediate, mac_address=fa:16:3e:42:f1:dd, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1636, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:30Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e142 e142: 6 total, 6 up, 6 in Nov 27 05:07:30 localhost dnsmasq[321799]: exiting on receipt of SIGTERM Nov 27 05:07:30 localhost podman[321835]: 2025-11-27 10:07:30.576549129 +0000 UTC m=+0.071904838 container kill 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:30 localhost systemd[1]: libpod-28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae.scope: Deactivated successfully. Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.581 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:30 localhost podman[321869]: 2025-11-27 10:07:30.666779248 +0000 UTC m=+0.063824730 container died 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:30 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:30 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:30 localhost podman[321882]: 2025-11-27 10:07:30.737383229 +0000 UTC m=+0.107616238 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:30 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.758 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.758 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.759 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.759 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:07:30 localhost nova_compute[284026]: 2025-11-27 10:07:30.759 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:07:30 localhost podman[321869]: 2025-11-27 10:07:30.765622139 +0000 UTC m=+0.162667621 container remove 28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:07:30 localhost systemd[1]: libpod-conmon-28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae.scope: Deactivated successfully. Nov 27 05:07:30 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:30.904 265123 INFO neutron.agent.dhcp.agent [None req-be47323c-6a47-4feb-875a-cf48b82a4176 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:30Z, description=, device_id=6937e43a-4a4c-4185-9e6e-85fe2023b27e, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=d9269413-6520-495a-a008-6e1d6b589fc6, ip_allocation=immediate, mac_address=fa:16:3e:97:a8:a5, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1639, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:30Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:30 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:30.967 2 INFO neutron.agent.securitygroups_rpc [None req-cb216a80-3af4-40cd-9f63-b7bca3523bf4 f3e59f86157c4a87bcd419c08b2737a3 199c995d53164d408ada510b321b5483 - - default default] Security group member updated ['2244145e-f58a-48f3-bccf-8e97f62f0a77']#033[00m Nov 27 05:07:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:31.074 265123 INFO neutron.agent.dhcp.agent [None req-5c58260c-5d81-4022-bd0c-1939aec78f1f - - - - - -] DHCP configuration for ports {'eaed8bc9-3c9b-4b50-a469-54f09e0b4dcc'} is completed#033[00m Nov 27 05:07:31 localhost systemd[1]: var-lib-containers-storage-overlay-4278585804dce584c080adf6f09d571c2f2a079678408ec63f158a50867a242f-merged.mount: Deactivated successfully. Nov 27 05:07:31 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-28959e1c7cd89676fc1396bfad7de58eebb09d49f88e4d93b41f78896b8d4fae-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:31 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:07:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:31 localhost podman[321946]: 2025-11-27 10:07:31.127995647 +0000 UTC m=+0.088499644 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:07:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v290: 177 pgs: 177 active+clean; 345 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 87 KiB/s rd, 25 MiB/s wr, 123 op/s Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.201 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:07:31 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/23760002' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.231 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.472s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.314 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.315 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:07:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:31.424 2 INFO neutron.agent.securitygroups_rpc [None req-06020dac-dc22-4840-ad26-0654e84c57d5 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:31.517 265123 INFO neutron.agent.dhcp.agent [None req-c262e15e-3701-49d6-9c29-fe416372ee40 - - - - - -] DHCP configuration for ports {'d9269413-6520-495a-a008-6e1d6b589fc6'} is completed#033[00m Nov 27 05:07:31 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:31 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:31 localhost podman[321986]: 2025-11-27 10:07:31.585469015 +0000 UTC m=+0.083531521 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.596 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.598 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11148MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.598 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:07:31 localhost nova_compute[284026]: 2025-11-27 10:07:31.598 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.022 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.022 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.023 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.358 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.362 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:32 localhost podman[322055]: Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.460 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.461 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 05:07:32 localhost podman[322055]: 2025-11-27 10:07:32.467821872 +0000 UTC m=+0.094866827 container create e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.481 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 05:07:32 localhost systemd[1]: Started libpod-conmon-e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac.scope. Nov 27 05:07:32 localhost podman[322055]: 2025-11-27 10:07:32.412780109 +0000 UTC m=+0.039825064 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:32 localhost systemd[1]: tmp-crun.DwA56m.mount: Deactivated successfully. Nov 27 05:07:32 localhost systemd[1]: Started libcrun container. Nov 27 05:07:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/ddf6f7794c547441176b83d4ef3bf79c1ed6bf0dc912cb6e739173dc4aaa0f13/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:32 localhost podman[322055]: 2025-11-27 10:07:32.536482382 +0000 UTC m=+0.163527328 container init e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2) Nov 27 05:07:32 localhost podman[322055]: 2025-11-27 10:07:32.545569187 +0000 UTC m=+0.172614152 container start e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:07:32 localhost dnsmasq[322073]: started, version 2.85 cachesize 150 Nov 27 05:07:32 localhost dnsmasq[322073]: DNS service limited to local subnets Nov 27 05:07:32 localhost dnsmasq[322073]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:32 localhost dnsmasq[322073]: warning: no upstream servers configured Nov 27 05:07:32 localhost dnsmasq-dhcp[322073]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:32 localhost dnsmasq[322073]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:32 localhost dnsmasq-dhcp[322073]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:32 localhost dnsmasq-dhcp[322073]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.651 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 05:07:32 localhost nova_compute[284026]: 2025-11-27 10:07:32.697 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:07:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e142 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:07:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3996521692' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:07:33 localhost nova_compute[284026]: 2025-11-27 10:07:33.195 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.498s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:07:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v291: 177 pgs: 177 active+clean; 345 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 75 KiB/s rd, 22 MiB/s wr, 106 op/s Nov 27 05:07:33 localhost nova_compute[284026]: 2025-11-27 10:07:33.202 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:07:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3528546135' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3528546135' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:33 localhost nova_compute[284026]: 2025-11-27 10:07:33.238 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:07:33 localhost nova_compute[284026]: 2025-11-27 10:07:33.239 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:07:33 localhost nova_compute[284026]: 2025-11-27 10:07:33.240 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.641s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:07:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:33.240 265123 INFO neutron.agent.dhcp.agent [None req-6c6e7619-b1be-4fa1-ac39-a256cdc08b76 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:33 localhost dnsmasq[322073]: exiting on receipt of SIGTERM Nov 27 05:07:33 localhost podman[322112]: 2025-11-27 10:07:33.322577448 +0000 UTC m=+0.043573503 container kill e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:33 localhost systemd[1]: libpod-e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac.scope: Deactivated successfully. Nov 27 05:07:33 localhost podman[322126]: 2025-11-27 10:07:33.410085394 +0000 UTC m=+0.067864418 container died e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:07:33 localhost systemd[1]: var-lib-containers-storage-overlay-ddf6f7794c547441176b83d4ef3bf79c1ed6bf0dc912cb6e739173dc4aaa0f13-merged.mount: Deactivated successfully. Nov 27 05:07:33 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:33 localhost podman[322126]: 2025-11-27 10:07:33.4786089 +0000 UTC m=+0.136387844 container cleanup e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:07:33 localhost systemd[1]: libpod-conmon-e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac.scope: Deactivated successfully. Nov 27 05:07:33 localhost podman[322127]: 2025-11-27 10:07:33.49980703 +0000 UTC m=+0.150442571 container remove e849b0117f6f827eaafd8cf4901727e01b6fec15edc2dd8cc69839ca215699ac (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:07:34 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:34 localhost podman[322199]: 2025-11-27 10:07:34.234506664 +0000 UTC m=+0.060886311 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:34 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:34 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:34 localhost podman[322237]: Nov 27 05:07:34 localhost podman[322237]: 2025-11-27 10:07:34.375826928 +0000 UTC m=+0.070495729 container create 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:07:34 localhost systemd[1]: Started libpod-conmon-793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5.scope. Nov 27 05:07:34 localhost systemd[1]: Started libcrun container. Nov 27 05:07:34 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/466eacf7481183dd858a7ec75c7f5f7dcd62911e73ff81f838a119d112646bb8/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:34 localhost podman[322237]: 2025-11-27 10:07:34.437504169 +0000 UTC m=+0.132173010 container init 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:07:34 localhost podman[322237]: 2025-11-27 10:07:34.444311673 +0000 UTC m=+0.138980514 container start 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:07:34 localhost podman[322237]: 2025-11-27 10:07:34.346221362 +0000 UTC m=+0.040890183 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:34 localhost dnsmasq[322261]: started, version 2.85 cachesize 150 Nov 27 05:07:34 localhost dnsmasq[322261]: DNS service limited to local subnets Nov 27 05:07:34 localhost dnsmasq[322261]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:34 localhost dnsmasq[322261]: warning: no upstream servers configured Nov 27 05:07:34 localhost dnsmasq-dhcp[322261]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:34 localhost dnsmasq[322261]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:34 localhost dnsmasq-dhcp[322261]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:34 localhost dnsmasq-dhcp[322261]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:34.739 265123 INFO neutron.agent.dhcp.agent [None req-3c980b08-1880-491e-a320-87eead815a1e - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:34 localhost podman[322279]: 2025-11-27 10:07:34.885190794 +0000 UTC m=+0.064469087 container kill 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:07:34 localhost dnsmasq[322261]: exiting on receipt of SIGTERM Nov 27 05:07:34 localhost systemd[1]: libpod-793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5.scope: Deactivated successfully. Nov 27 05:07:34 localhost podman[322293]: 2025-11-27 10:07:34.979486193 +0000 UTC m=+0.068200108 container died 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:35 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:35 localhost systemd[1]: var-lib-containers-storage-overlay-466eacf7481183dd858a7ec75c7f5f7dcd62911e73ff81f838a119d112646bb8-merged.mount: Deactivated successfully. Nov 27 05:07:35 localhost podman[322293]: 2025-11-27 10:07:35.02918268 +0000 UTC m=+0.117896565 container remove 793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:07:35 localhost systemd[1]: libpod-conmon-793a4db2e16928bfdb874d4a5218ba199efa2827fe7f5e456da0d0685d2171b5.scope: Deactivated successfully. Nov 27 05:07:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v292: 177 pgs: 177 active+clean; 345 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 70 KiB/s rd, 20 MiB/s wr, 98 op/s Nov 27 05:07:35 localhost nova_compute[284026]: 2025-11-27 10:07:35.584 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:35 localhost podman[322367]: Nov 27 05:07:35 localhost podman[322367]: 2025-11-27 10:07:35.895633091 +0000 UTC m=+0.075283368 container create af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:07:35 localhost systemd[1]: Started libpod-conmon-af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01.scope. Nov 27 05:07:35 localhost systemd[1]: Started libcrun container. Nov 27 05:07:35 localhost podman[322367]: 2025-11-27 10:07:35.857221087 +0000 UTC m=+0.036871454 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:35 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/55a04a0a8c513742dd4342fe8683d77760af03a14a489ef007a0c0b7476f849b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:35 localhost podman[322367]: 2025-11-27 10:07:35.988300236 +0000 UTC m=+0.167950553 container init af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:35 localhost podman[322367]: 2025-11-27 10:07:35.999701703 +0000 UTC m=+0.179352020 container start af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:07:36 localhost dnsmasq[322385]: started, version 2.85 cachesize 150 Nov 27 05:07:36 localhost dnsmasq[322385]: DNS service limited to local subnets Nov 27 05:07:36 localhost dnsmasq[322385]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:36 localhost dnsmasq[322385]: warning: no upstream servers configured Nov 27 05:07:36 localhost dnsmasq-dhcp[322385]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:36 localhost dnsmasq[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:36 localhost dnsmasq-dhcp[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:36 localhost dnsmasq-dhcp[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:36.146 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:35Z, description=, device_id=81f150e2-9d47-42e2-9b0b-968902160e4b, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=37d51988-9367-4498-999c-b09f2510584a, ip_allocation=immediate, mac_address=fa:16:3e:e0:cb:77, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1652, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:35Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.204 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.240 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.242 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.242 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:07:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:36.304 265123 INFO neutron.agent.dhcp.agent [None req-25d84fa1-1f76-4154-a27c-a43ed7f65f79 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:36 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:36 localhost podman[322424]: 2025-11-27 10:07:36.368616537 +0000 UTC m=+0.060596763 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.386 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.386 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.386 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:07:36 localhost nova_compute[284026]: 2025-11-27 10:07:36.386 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:07:36 localhost dnsmasq[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:36 localhost dnsmasq-dhcp[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:36 localhost podman[322414]: 2025-11-27 10:07:36.425320384 +0000 UTC m=+0.142897759 container kill af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:07:36 localhost dnsmasq-dhcp[322385]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:36.892 265123 INFO neutron.agent.dhcp.agent [None req-d394afb5-bb5c-483d-8a8e-8ac2b17a3d2f - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea', '37d51988-9367-4498-999c-b09f2510584a'} is completed#033[00m Nov 27 05:07:36 localhost dnsmasq[322385]: exiting on receipt of SIGTERM Nov 27 05:07:36 localhost podman[322477]: 2025-11-27 10:07:36.921392761 +0000 UTC m=+0.058673971 container kill af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:07:36 localhost systemd[1]: libpod-af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01.scope: Deactivated successfully. Nov 27 05:07:37 localhost podman[322489]: 2025-11-27 10:07:37.001707544 +0000 UTC m=+0.067872420 container died af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:07:37 localhost systemd[1]: tmp-crun.SKblSC.mount: Deactivated successfully. Nov 27 05:07:37 localhost podman[322489]: 2025-11-27 10:07:37.041176726 +0000 UTC m=+0.107341552 container cleanup af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2) Nov 27 05:07:37 localhost systemd[1]: libpod-conmon-af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01.scope: Deactivated successfully. Nov 27 05:07:37 localhost podman[322491]: 2025-11-27 10:07:37.08029159 +0000 UTC m=+0.137342300 container remove af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 05:07:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v293: 177 pgs: 177 active+clean; 529 MiB data, 1.9 GiB used, 40 GiB / 42 GiB avail; 62 KiB/s rd, 33 MiB/s wr, 91 op/s Nov 27 05:07:37 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:37.288 2 INFO neutron.agent.securitygroups_rpc [None req-670a2d94-f336-4a2e-9d9f-44f74a5a8eff 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.513 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.666 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.695 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.696 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:37 localhost nova_compute[284026]: 2025-11-27 10:07:37.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 05:07:37 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:37.792 2 INFO neutron.agent.securitygroups_rpc [None req-8c29a691-835f-41dc-b690-6b6931a7f813 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e142 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:37 localhost systemd[1]: var-lib-containers-storage-overlay-55a04a0a8c513742dd4342fe8683d77760af03a14a489ef007a0c0b7476f849b-merged.mount: Deactivated successfully. Nov 27 05:07:37 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-af8047c7d3bf7a6124f0757ab4be92617b2343535113b617a6f5c8b554e5bb01-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:38 localhost podman[322569]: Nov 27 05:07:38 localhost podman[322569]: 2025-11-27 10:07:38.515118754 +0000 UTC m=+0.087844167 container create 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:07:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:07:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:07:38 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:07:38 localhost systemd[1]: Started libpod-conmon-5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330.scope. Nov 27 05:07:38 localhost podman[322569]: 2025-11-27 10:07:38.472189537 +0000 UTC m=+0.044915000 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:38 localhost systemd[1]: Started libcrun container. Nov 27 05:07:38 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/e20cabf896b6aee77950fe9986bc96d0d2b3a6f0cef9190e47e28e608524a3e8/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:38 localhost podman[322569]: 2025-11-27 10:07:38.596245628 +0000 UTC m=+0.168971101 container init 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:07:38 localhost podman[322569]: 2025-11-27 10:07:38.609405312 +0000 UTC m=+0.182130715 container start 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:38 localhost dnsmasq[322615]: started, version 2.85 cachesize 150 Nov 27 05:07:38 localhost dnsmasq[322615]: DNS service limited to local subnets Nov 27 05:07:38 localhost dnsmasq[322615]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:38 localhost dnsmasq[322615]: warning: no upstream servers configured Nov 27 05:07:38 localhost dnsmasq-dhcp[322615]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:38 localhost dnsmasq-dhcp[322615]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:38 localhost dnsmasq[322615]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:38 localhost dnsmasq-dhcp[322615]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:38 localhost dnsmasq-dhcp[322615]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:38 localhost podman[322582]: 2025-11-27 10:07:38.650852138 +0000 UTC m=+0.090747494 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:07:38 localhost podman[322582]: 2025-11-27 10:07:38.664314101 +0000 UTC m=+0.104209477 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, managed_by=edpm_ansible, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, config_id=edpm) Nov 27 05:07:38 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:07:38 localhost podman[322583]: 2025-11-27 10:07:38.718980353 +0000 UTC m=+0.155799686 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:07:38 localhost podman[322583]: 2025-11-27 10:07:38.730901684 +0000 UTC m=+0.167721007 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:07:38 localhost nova_compute[284026]: 2025-11-27 10:07:38.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:38 localhost nova_compute[284026]: 2025-11-27 10:07:38.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 05:07:38 localhost podman[242678]: time="2025-11-27T10:07:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:07:38 localhost nova_compute[284026]: 2025-11-27 10:07:38.761 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 05:07:38 localhost nova_compute[284026]: 2025-11-27 10:07:38.762 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:38 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:07:38 localhost podman[322584]: 2025-11-27 10:07:38.770834659 +0000 UTC m=+0.205093203 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, distribution-scope=public, build-date=2025-08-20T13:12:41, io.openshift.expose-services=, config_id=edpm, name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., version=9.6, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.buildah.version=1.33.7, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, release=1755695350, vendor=Red Hat, Inc., container_name=openstack_network_exporter, vcs-type=git, com.redhat.component=ubi9-minimal-container) Nov 27 05:07:38 localhost podman[242678]: @ - - [27/Nov/2025:10:07:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159424 "" "Go-http-client/1.1" Nov 27 05:07:38 localhost podman[322584]: 2025-11-27 10:07:38.855910589 +0000 UTC m=+0.290169163 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, vendor=Red Hat, Inc., io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, vcs-type=git, distribution-scope=public, name=ubi9-minimal, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, container_name=openstack_network_exporter, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc.) Nov 27 05:07:38 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:07:38 localhost systemd[1]: tmp-crun.dlYzuz.mount: Deactivated successfully. Nov 27 05:07:38 localhost podman[242678]: @ - - [27/Nov/2025:10:07:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20204 "" "Go-http-client/1.1" Nov 27 05:07:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:38.940 265123 INFO neutron.agent.dhcp.agent [None req-535c0041-8a24-42f3-a3f6-479a48e0963f - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:39 localhost dnsmasq[322615]: exiting on receipt of SIGTERM Nov 27 05:07:39 localhost podman[322664]: 2025-11-27 10:07:39.067764474 +0000 UTC m=+0.055595978 container kill 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:39 localhost systemd[1]: libpod-5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330.scope: Deactivated successfully. Nov 27 05:07:39 localhost podman[322675]: 2025-11-27 10:07:39.143181705 +0000 UTC m=+0.060962153 container died 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v294: 177 pgs: 177 active+clean; 529 MiB data, 1.9 GiB used, 40 GiB / 42 GiB avail; 62 KiB/s rd, 33 MiB/s wr, 91 op/s Nov 27 05:07:39 localhost podman[322675]: 2025-11-27 10:07:39.244348569 +0000 UTC m=+0.162128987 container cleanup 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:39 localhost systemd[1]: libpod-conmon-5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330.scope: Deactivated successfully. Nov 27 05:07:39 localhost podman[322677]: 2025-11-27 10:07:39.264424059 +0000 UTC m=+0.174793287 container remove 5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:07:39 localhost systemd[1]: var-lib-containers-storage-overlay-e20cabf896b6aee77950fe9986bc96d0d2b3a6f0cef9190e47e28e608524a3e8-merged.mount: Deactivated successfully. Nov 27 05:07:39 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-5f4b5743ceccc01234bbf35b07b6a8fa961f67b14855bc3a4ce5a5605191b330-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:40 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:40 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:40 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:40 localhost podman[322749]: 2025-11-27 10:07:40.098851677 +0000 UTC m=+0.115388828 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:07:40 localhost podman[322780]: Nov 27 05:07:40 localhost podman[322780]: 2025-11-27 10:07:40.198923062 +0000 UTC m=+0.141062580 container create 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:40 localhost podman[322780]: 2025-11-27 10:07:40.13643914 +0000 UTC m=+0.078578638 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:40 localhost systemd[1]: Started libpod-conmon-32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e.scope. Nov 27 05:07:40 localhost systemd[1]: Started libcrun container. Nov 27 05:07:40 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/60e154cfb7573fa5b6e5b1df0e3c2ccdc9086e1531ff70f746f37782b1618707/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:40 localhost podman[322780]: 2025-11-27 10:07:40.288978167 +0000 UTC m=+0.231117665 container init 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:07:40 localhost podman[322780]: 2025-11-27 10:07:40.295680767 +0000 UTC m=+0.237820265 container start 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:40 localhost dnsmasq[322807]: started, version 2.85 cachesize 150 Nov 27 05:07:40 localhost dnsmasq[322807]: DNS service limited to local subnets Nov 27 05:07:40 localhost dnsmasq[322807]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:40 localhost dnsmasq[322807]: warning: no upstream servers configured Nov 27 05:07:40 localhost dnsmasq-dhcp[322807]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:40 localhost dnsmasq[322807]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:40 localhost dnsmasq-dhcp[322807]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:40 localhost dnsmasq-dhcp[322807]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:40 localhost nova_compute[284026]: 2025-11-27 10:07:40.587 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:40 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:40.816 265123 INFO neutron.agent.dhcp.agent [None req-9f58f34e-0dcf-49d9-a402-15361deabe6e - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:40 localhost dnsmasq[322807]: exiting on receipt of SIGTERM Nov 27 05:07:40 localhost podman[322825]: 2025-11-27 10:07:40.821199237 +0000 UTC m=+0.063488461 container kill 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:40 localhost systemd[1]: libpod-32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e.scope: Deactivated successfully. Nov 27 05:07:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:40 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3372220366' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:40 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3372220366' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:40 localhost podman[322841]: 2025-11-27 10:07:40.899174897 +0000 UTC m=+0.056885794 container died 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:07:40 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:40 localhost systemd[1]: var-lib-containers-storage-overlay-60e154cfb7573fa5b6e5b1df0e3c2ccdc9086e1531ff70f746f37782b1618707-merged.mount: Deactivated successfully. Nov 27 05:07:40 localhost podman[322841]: 2025-11-27 10:07:40.997946356 +0000 UTC m=+0.155657203 container remove 32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:41 localhost systemd[1]: libpod-conmon-32715f9c1dbf3654a0a65ee154e5f594cd79ae3a30a77f08244f8d9614bca76e.scope: Deactivated successfully. Nov 27 05:07:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v295: 177 pgs: 177 active+clean; 713 MiB data, 2.4 GiB used, 40 GiB / 42 GiB avail; 49 KiB/s rd, 34 MiB/s wr, 75 op/s Nov 27 05:07:41 localhost nova_compute[284026]: 2025-11-27 10:07:41.208 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:41 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:41.634 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:41Z, description=, device_id=dcc5b01e-054a-464b-9a9a-83608c157eb1, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=ea180b72-46cc-4b7e-8557-ffebae17669c, ip_allocation=immediate, mac_address=fa:16:3e:e1:5c:a9, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1673, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:41Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:41 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:41 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:41 localhost podman[322891]: 2025-11-27 10:07:41.870743128 +0000 UTC m=+0.067044267 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:07:41 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:42.075 265123 INFO neutron.agent.dhcp.agent [None req-287fcdc9-9866-4472-b099-0be25e4d251d - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:41Z, description=, device_id=5f5918b1-51d3-4b84-b4d9-6df3779a1472, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=28e2c199-0051-47da-937d-06c2efdc5250, ip_allocation=immediate, mac_address=fa:16:3e:2d:78:6c, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1674, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:41Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:42.198 265123 INFO neutron.agent.dhcp.agent [None req-9d414c33-9dea-4620-8a72-0a9b69376a60 - - - - - -] DHCP configuration for ports {'ea180b72-46cc-4b7e-8557-ffebae17669c'} is completed#033[00m Nov 27 05:07:42 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 5 addresses Nov 27 05:07:42 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:42 localhost podman[322953]: 2025-11-27 10:07:42.324394733 +0000 UTC m=+0.064042566 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:42 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e143 e143: 6 total, 6 up, 6 in Nov 27 05:07:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:42.631 265123 INFO neutron.agent.dhcp.agent [None req-67637c66-c9e5-49a4-8a32-3e6bfa299e46 - - - - - -] DHCP configuration for ports {'28e2c199-0051-47da-937d-06c2efdc5250'} is completed#033[00m Nov 27 05:07:42 localhost podman[322995]: Nov 27 05:07:42 localhost podman[322995]: 2025-11-27 10:07:42.663410511 +0000 UTC m=+0.092165653 container create d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:42 localhost systemd[1]: Started libpod-conmon-d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff.scope. Nov 27 05:07:42 localhost systemd[1]: Started libcrun container. Nov 27 05:07:42 localhost podman[322995]: 2025-11-27 10:07:42.620435993 +0000 UTC m=+0.049191185 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:42 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/5e21b422954d9fbc31b1f8ffb1b887ff998664bfffb387ff148ba99ce1a41fcc/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:42 localhost podman[322995]: 2025-11-27 10:07:42.738399429 +0000 UTC m=+0.167154571 container init d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:42 localhost podman[322995]: 2025-11-27 10:07:42.747009811 +0000 UTC m=+0.175764963 container start d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:07:42 localhost dnsmasq[323013]: started, version 2.85 cachesize 150 Nov 27 05:07:42 localhost dnsmasq[323013]: DNS service limited to local subnets Nov 27 05:07:42 localhost dnsmasq[323013]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:42 localhost dnsmasq[323013]: warning: no upstream servers configured Nov 27 05:07:42 localhost dnsmasq-dhcp[323013]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:42 localhost dnsmasq[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:42 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:42 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:42 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:42.810 265123 INFO neutron.agent.dhcp.agent [None req-fff1820d-4822-42ad-b72b-4d5f9ce224fa - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:29Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=5d4cdb4b-456d-464a-b316-7371c5e007a8, ip_allocation=immediate, mac_address=fa:16:3e:4a:a4:48, name=tempest-NetworksTestDHCPv6-824830827, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=49, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['17d52702-5eb1-4f69-879f-72ccfef606fe', '384debcb-5666-43a8-a812-56889f8732c9'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:28Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1635, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:30Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e143 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:43 localhost dnsmasq[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:07:43 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:43 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:43 localhost podman[323029]: 2025-11-27 10:07:43.014646458 +0000 UTC m=+0.063472480 container kill d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:43.021 265123 INFO neutron.agent.dhcp.agent [None req-acfc612e-8157-4399-8862-d5079e246752 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:43.155 265123 INFO neutron.agent.dhcp.agent [None req-fff1820d-4822-42ad-b72b-4d5f9ce224fa - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:36Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=396a82cb-4a46-4ec8-bfb4-5370135ba06b, ip_allocation=immediate, mac_address=fa:16:3e:12:11:9d, name=tempest-NetworksTestDHCPv6-1434689652, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=53, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['211e35f5-16f4-40ff-abe3-19361896ccb3', 'a385e33f-de04-4f45-9a1f-12d50964d47a'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:36Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1654, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:37Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v297: 177 pgs: 177 active+clean; 713 MiB data, 2.4 GiB used, 40 GiB / 42 GiB avail; 52 KiB/s rd, 37 MiB/s wr, 80 op/s Nov 27 05:07:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:43.256 265123 INFO neutron.agent.dhcp.agent [None req-6f70dd9b-f89c-499a-81f6-88f84e3f47ec - - - - - -] DHCP configuration for ports {'5d4cdb4b-456d-464a-b316-7371c5e007a8'} is completed#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.330 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:37:8a:8d 2001:db8:0:1:f816:3eff:fe37:8a8d'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:0:1:f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '30', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=2e1b06ae-5268-4e92-89b9-6d951a8114ea) old=Port_Binding(mac=['fa:16:3e:37:8a:8d 2001:db8::f816:3eff:fe37:8a8d'], external_ids={'neutron:cidrs': '2001:db8::f816:3eff:fe37:8a8d/64', 'neutron:device_id': 'ovnmeta-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '28', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.332 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port 2e1b06ae-5268-4e92-89b9-6d951a8114ea in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd updated#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.334 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port ed52eec8-9f5d-45db-b72d-a5c9e33828a3 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.334 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.336 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[17eafe66-020d-4453-a7aa-ccd5cdbe4bac]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:43 localhost dnsmasq[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 4 addresses Nov 27 05:07:43 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:43 localhost systemd[1]: tmp-crun.MjuBXC.mount: Deactivated successfully. Nov 27 05:07:43 localhost dnsmasq-dhcp[323013]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:43 localhost podman[323067]: 2025-11-27 10:07:43.350308456 +0000 UTC m=+0.075373320 container kill d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:07:43 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:43 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:43 localhost podman[323101]: 2025-11-27 10:07:43.539095609 +0000 UTC m=+0.073829318 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:43 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.576 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.577 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:07:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:43.577 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:07:43 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:43.588 265123 INFO neutron.agent.dhcp.agent [None req-d67236c7-5fa0-4d3e-bf50-02ff48f0ee8c - - - - - -] DHCP configuration for ports {'396a82cb-4a46-4ec8-bfb4-5370135ba06b'} is completed#033[00m Nov 27 05:07:43 localhost dnsmasq[323013]: exiting on receipt of SIGTERM Nov 27 05:07:43 localhost podman[323143]: 2025-11-27 10:07:43.797125467 +0000 UTC m=+0.052616868 container kill d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:07:43 localhost systemd[1]: libpod-d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff.scope: Deactivated successfully. Nov 27 05:07:43 localhost podman[323158]: 2025-11-27 10:07:43.854809741 +0000 UTC m=+0.043578585 container died d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:07:43 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:43 localhost systemd[1]: var-lib-containers-storage-overlay-5e21b422954d9fbc31b1f8ffb1b887ff998664bfffb387ff148ba99ce1a41fcc-merged.mount: Deactivated successfully. Nov 27 05:07:43 localhost podman[323158]: 2025-11-27 10:07:43.928143135 +0000 UTC m=+0.116911949 container cleanup d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:07:43 localhost systemd[1]: libpod-conmon-d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff.scope: Deactivated successfully. Nov 27 05:07:43 localhost ovn_controller[156436]: 2025-11-27T10:07:43Z|00245|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:07:43 localhost podman[323159]: 2025-11-27 10:07:43.947823355 +0000 UTC m=+0.128123211 container remove d20da2d1ef5e62ec91d54b36388c0cf9732e6e69a4fd1a43c2eaab3215a370ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:43 localhost nova_compute[284026]: 2025-11-27 10:07:43.990 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.138 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.138 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.170 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2605e037-e66f-4f9d-99b7-6dd0ac4d7aad', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.139201', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaa54dc8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': 'fc60452036ec06e99311fabc6d6c22df89d4a8aae450ca2be4ef0833c129ad35'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.139201', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaa5640c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': 'c762ebf2edcbc185efdb7d89e663f2772c48aa9eb26ff1212b5248a4b69ff252'}]}, 'timestamp': '2025-11-27 10:07:44.170919', '_unique_id': '7ffd6aea10bc404d90b02d8fb836befc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.172 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.173 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.174 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.174 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.174 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c528aa99-d04c-423f-b919-18d5c02a6c03', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.174172', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaa5f430-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '919c671755b7da4b70a9c7ea6148fbc51d5103c5bfac7777feba4a67b9f62a54'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.174172', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaa6054c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '9ebbe32da3394b72ffdb695de774c77984ca04c96e433d09ca08e8c10eb57365'}]}, 'timestamp': '2025-11-27 10:07:44.175028', '_unique_id': '1d62187b69fd4e06a490b79f37f0cdb7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.175 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.177 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.177 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.181 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '74f58507-f179-463e-a04c-009f659af815', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.177298', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaa719dc-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '4687fed5e32e52a4d1155f986aa8c11db2005235e0f987c9808d3c5ed4ea54c0'}]}, 'timestamp': '2025-11-27 10:07:44.182151', '_unique_id': '40a194deb7f547b4ae831d2bbebfab8c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.184 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.200 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.200 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '069344cd-8ce5-4c07-b7e4-de792aa2bda4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.184296', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaa9f99a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': '3ca5bd31c554f84598979731ecfa1dfb46acd3fe7a9065e4589efcbe47d3b283'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.184296', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaaa0a02-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': 'f56bb6bfd5bb8538692063f5b697c1918992e9668dda8366585f0f500793b379'}]}, 'timestamp': '2025-11-27 10:07:44.201388', '_unique_id': '0ac858850e3e41a79f0dc2c457331ec6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.202 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.203 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.203 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f0058c71-b30a-4dc0-ad0e-625ef5ffacc4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.203705', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaaa7618-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '71d63ee5c24f6b2507a996765f1f092d75f6d3faf1c5f7ae27795598a1696362'}]}, 'timestamp': '2025-11-27 10:07:44.204165', '_unique_id': 'd0859dda4c9b41c2a551daef2f28dc62'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.205 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.206 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.206 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.206 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3f916281-ce07-4ea0-8c1e-8f55ebac6950', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.206390', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaaae04e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': 'd1d3ce9a36afb33ed92f11451345f8c2b1fa32f3c2f5a3e35401ea6675a12152'}]}, 'timestamp': '2025-11-27 10:07:44.206879', '_unique_id': 'd9219458f28d4215a068b67330cf66e2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.207 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.208 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.209 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '41bf9f21-d1e9-4822-999e-57d3950030f7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.208974', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaab4412-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '5157ff2ebb687f0940c8bfb489a35ef294e62c5d534269912db1f17f94b3dccd'}]}, 'timestamp': '2025-11-27 10:07:44.209435', '_unique_id': '50337a77834b43d5a6918a3b36d88c92'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.210 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.211 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd89edf7a-8d1e-4cd5-8936-7e4a59f686f4', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.211467', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaaba682-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': 'd73bfc1912e137a434426d7fdee428766b1ad00639c9c43143487b7bb2cf9c08'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.211467', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaabb5dc-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '219c34754c76ccd2922d85b92ea00f427209bae333f0a6c60e00e2a04f848ecc'}]}, 'timestamp': '2025-11-27 10:07:44.212315', '_unique_id': 'e28d188ddd254b1d9b7e4bce20004fe9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.214 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '81be3170-cf76-418b-8b5f-00cf8f3da776', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.214572', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaac1e6e-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '49b5be44cb123292b1e6344ce7c4b8798ca42f1683185fceb231efa7117c95a6'}]}, 'timestamp': '2025-11-27 10:07:44.215022', '_unique_id': '6db12d47d83d45a690d89a9ff85e74cc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.215 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.216 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '315905e8-8f7b-423b-8295-a36103f138e6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.217088', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaac8020-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '3cda001240006e4ec850a9ccc3d5da1427fd1a89d67b95fc92d1ecfbe4b30c22'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.217088', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaac90c4-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '22af82efb5fe215d944d5d37e25f1cc464f9a3f53b254bab9b79a29785f06145'}]}, 'timestamp': '2025-11-27 10:07:44.217921', '_unique_id': 'cd5d57cc6d7647ba9c3732d0f5783e3c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.219 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c25caddc-11b3-408b-bb9c-3216c3ee9472', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.220001', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaacf208-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '413c0652c95c6541c238fdaf81fc32bf31d17a4c458d1e9e1298bd6c02147624'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.220001', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaad0284-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '14a64603c07496a8e3fb897674dfcdf21882072f5b2660e9fbc90db151ca3197'}]}, 'timestamp': '2025-11-27 10:07:44.220833', '_unique_id': '358f8b3598fa45afbf9eee68285431d7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.221 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.222 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0460ed61-0f46-4f67-8073-55b0e20836f6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.223080', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eaad6a80-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': '17abcec6e1fbbbce5cb26334e53468c419cb39ad99a466a979be0d52f6d0978c'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.223080', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eaad7c00-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': '5850a41b44acee1e36dea8fbb8b62963687801cf0a58e3e8d8f211ea70b17fff'}]}, 'timestamp': '2025-11-27 10:07:44.223943', '_unique_id': 'c73ebaf7cf9641c3a9ea3e651491b24a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.224 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.225 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b2222724-fd8e-4c14-886c-a5fbf401ab9a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.226051', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eaaddb8c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': 'f0e9a6b7aa8f184b5313d2c0280de724ea16e118c2f1f24cc012902aca54d252'}]}, 'timestamp': '2025-11-27 10:07:44.226331', '_unique_id': '5c350380830f4591a3bc9ec9f1690fd4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.226 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.227 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 16730000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9e57fb00-8061-463d-8242-27cb59684a0b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 16730000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:07:44.227632', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'eab09a0c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.431980292, 'message_signature': '67560cc7228706ee5423da9258d75d83baa7f472a47447f52a29202e47cf7f78'}]}, 'timestamp': '2025-11-27 10:07:44.244311', '_unique_id': 'c3e0a249b4c74e0a8fd649a3d419b17a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c4e43313-9733-4b00-8097-22aff19d9a08', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.245681', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eab0da6c-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': 'b672f29e64fda627bece59448eed781acef0a6d7c1a8aa702a6b14ce7e8d4db9'}]}, 'timestamp': '2025-11-27 10:07:44.245965', '_unique_id': 'baf7445b93204c3683bef28664b79c07'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b634c54a-6427-4e91-838a-fef3e9320450', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.247259', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eab117fc-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '6de69c5c54fde0a13d0244940de4ea5912883226bbc0dd54b4c2993af8c9e139'}]}, 'timestamp': '2025-11-27 10:07:44.247562', '_unique_id': 'bc4cf2387e3543b4a553df8dccf5f6c6'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'afcef8a4-8f62-4cf9-8c03-4c5f8c1b755d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.248870', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eab156b8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': '7046bf03d20aa51f2e16ae1d578e321f121e85fa1af01eb465cc67a45da8bd1d'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.248870', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eab1609a-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.372349746, 'message_signature': '2f511dacd9b8b2deb98a985b2cdbcff63eb3f9af7828cf69cc76bcbcc00463f5'}]}, 'timestamp': '2025-11-27 10:07:44.249382', '_unique_id': 'b045dcaa6014436e9767e58d3c242b6a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'baf39568-689f-441a-9d53-c284fea57285', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.250717', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eab19f24-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': '854af2f77dc82c175bb1bbf800f7cc25f206b9b09a9b2800c9b608cdb1eff203'}]}, 'timestamp': '2025-11-27 10:07:44.251000', '_unique_id': '2ff33ba5cf34431db8051f0c9fb1f3d5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.252 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.252 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '43d7d98b-eff5-43ec-b41e-04cb5b3c4fc5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:07:44.252551', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'eab1e6c8-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.365372348, 'message_signature': 'b3fdd9886b73c9bf96853c2bd435d85dd21055f491b8960ead07ef8cd37b06ac'}]}, 'timestamp': '2025-11-27 10:07:44.252832', '_unique_id': 'b0b68656ea8b470faade895c46d82a79'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.253 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.254 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.254 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.254 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1336e0b7-a0fa-4c33-8b3b-691cda88fd95', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:07:44.254100', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'eab22322-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '90691f561c2e1500f5266a28f5f90dfd8034653c935ed4c32efb250252dd07ed'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:07:44.254100', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'eab22cfa-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.327255131, 'message_signature': '8ab9db21d2161f62655be56fdc414e53d7b820c19428967f42bb54d025d17e86'}]}, 'timestamp': '2025-11-27 10:07:44.254634', '_unique_id': '8a1a0a86fe2d4a43a5037793733d5d4b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1b43387c-7f62-4977-9ad3-0a53540b8428', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:07:44.255939', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'eab26b02-cb78-11f0-8166-fa163e6bfee1', 'monotonic_time': 12294.431980292, 'message_signature': '48c999e22a47c0ea96b11e36297329137a5a9d01c2c3a7fd3b86d5a37b827b1f'}]}, 'timestamp': '2025-11-27 10:07:44.256207', '_unique_id': '77b7d1271b054a588eb3a36a654e97ab'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:07:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:07:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:07:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:07:44 Nov 27 05:07:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:07:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:07:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['volumes', 'vms', 'images', '.mgr', 'backups', 'manila_metadata', 'manila_data'] Nov 27 05:07:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:07:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e144 e144: 6 total, 6 up, 6 in Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:07:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:07:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:44.708 2 INFO neutron.agent.securitygroups_rpc [None req-c8d3c13e-01f6-49a2-833f-aca4f25edb5e 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 2.9989356504745952e-06 of space, bias 1.0, pg target 0.0005987874848780942 quantized to 32 (current 32) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.04394258619486321 of space, bias 1.0, pg target 8.773869710241021 quantized to 32 (current 32) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:07:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002082897124511446 quantized to 16 (current 16) Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:07:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:07:45 localhost podman[323237]: Nov 27 05:07:45 localhost podman[323237]: 2025-11-27 10:07:45.024159926 +0000 UTC m=+0.117366231 container create 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:07:45 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:45.041 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=16, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=15) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:45 localhost nova_compute[284026]: 2025-11-27 10:07:45.042 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:45 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:45.043 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 5 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:07:45 localhost podman[323237]: 2025-11-27 10:07:44.959984078 +0000 UTC m=+0.053190393 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:45 localhost systemd[1]: Started libpod-conmon-894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870.scope. Nov 27 05:07:45 localhost systemd[1]: Started libcrun container. Nov 27 05:07:45 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/c1397f090517c53aef071372b627f0144db1f94ff70c60564f598485d0c7b7e1/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:45 localhost podman[323237]: 2025-11-27 10:07:45.119115513 +0000 UTC m=+0.212321818 container init 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:07:45 localhost podman[323237]: 2025-11-27 10:07:45.127993272 +0000 UTC m=+0.221199627 container start 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:45 localhost dnsmasq[323255]: started, version 2.85 cachesize 150 Nov 27 05:07:45 localhost dnsmasq[323255]: DNS service limited to local subnets Nov 27 05:07:45 localhost dnsmasq[323255]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:45 localhost dnsmasq[323255]: warning: no upstream servers configured Nov 27 05:07:45 localhost dnsmasq-dhcp[323255]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:45 localhost dnsmasq[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:45 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:45 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:45.185 265123 INFO neutron.agent.dhcp.agent [None req-0dc17f9d-4b88-46f5-a9d5-9c340e20acda - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:44Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=a70d36e4-80df-4cb5-8ad1-0968d24d48ab, ip_allocation=immediate, mac_address=fa:16:3e:83:02:60, name=tempest-NetworksTestDHCPv6-237696355, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=57, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['41121fa8-b158-4348-ada3-6408a88b4d0a', '9681f90c-ae28-4776-8ef8-fd116bf3e8ed'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:40Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1680, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:44Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v299: 177 pgs: 177 active+clean; 713 MiB data, 2.4 GiB used, 40 GiB / 42 GiB avail; 32 KiB/s rd, 23 MiB/s wr, 50 op/s Nov 27 05:07:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:45 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3162869913' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:45 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3162869913' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:45.340 265123 INFO neutron.agent.dhcp.agent [None req-5a2fd683-9a12-4609-8e4e-380662846602 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:45 localhost dnsmasq[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:07:45 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:45 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:45 localhost podman[323274]: 2025-11-27 10:07:45.396637896 +0000 UTC m=+0.056968575 container kill 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:07:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e145 e145: 6 total, 6 up, 6 in Nov 27 05:07:45 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:45 localhost podman[323305]: 2025-11-27 10:07:45.565900924 +0000 UTC m=+0.074117247 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:45 localhost nova_compute[284026]: 2025-11-27 10:07:45.590 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:45 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:45.697 265123 INFO neutron.agent.dhcp.agent [None req-5c1545ba-cab2-43e4-be91-c64750673654 - - - - - -] DHCP configuration for ports {'a70d36e4-80df-4cb5-8ad1-0968d24d48ab'} is completed#033[00m Nov 27 05:07:45 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:45.858 2 INFO neutron.agent.securitygroups_rpc [None req-9ec7ed64-029b-4e7c-9aa8-34474b619f7d 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:07:46 localhost podman[323333]: 2025-11-27 10:07:46.018080439 +0000 UTC m=+0.104367941 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:07:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:07:46 localhost podman[323333]: 2025-11-27 10:07:46.138163612 +0000 UTC m=+0.224451104 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:07:46 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:07:46 localhost nova_compute[284026]: 2025-11-27 10:07:46.211 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:46 localhost systemd[1]: tmp-crun.EmRiII.mount: Deactivated successfully. Nov 27 05:07:46 localhost podman[323368]: 2025-11-27 10:07:46.220244362 +0000 UTC m=+0.166418311 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:46 localhost podman[323368]: 2025-11-27 10:07:46.285083828 +0000 UTC m=+0.231257787 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.vendor=CentOS, config_id=ovn_controller, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, container_name=ovn_controller, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:46 localhost dnsmasq[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:46 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:46 localhost podman[323382]: 2025-11-27 10:07:46.287287527 +0000 UTC m=+0.167225344 container kill 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:46 localhost dnsmasq-dhcp[323255]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:46 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:07:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:46.728 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:46Z, description=, device_id=7c30f5dd-4f2a-4e5d-ab18-5b625f70e7a9, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=5b11b185-fbb2-4d32-8799-c379f25a6bbb, ip_allocation=immediate, mac_address=fa:16:3e:02:12:b0, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1683, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:46Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:47 localhost podman[323443]: 2025-11-27 10:07:47.013418769 +0000 UTC m=+0.110241069 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:07:47 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:47 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:47 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:47 localhost dnsmasq[323255]: exiting on receipt of SIGTERM Nov 27 05:07:47 localhost podman[323462]: 2025-11-27 10:07:47.046065338 +0000 UTC m=+0.062930595 container kill 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:07:47 localhost systemd[1]: libpod-894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870.scope: Deactivated successfully. Nov 27 05:07:47 localhost podman[323478]: 2025-11-27 10:07:47.111288204 +0000 UTC m=+0.046223965 container died 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:07:47 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:47 localhost systemd[1]: var-lib-containers-storage-overlay-c1397f090517c53aef071372b627f0144db1f94ff70c60564f598485d0c7b7e1-merged.mount: Deactivated successfully. Nov 27 05:07:47 localhost podman[323478]: 2025-11-27 10:07:47.199573492 +0000 UTC m=+0.134509233 container remove 894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:07:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v301: 177 pgs: 177 active+clean; 889 MiB data, 2.9 GiB used, 39 GiB / 42 GiB avail; 54 KiB/s rd, 29 MiB/s wr, 83 op/s Nov 27 05:07:47 localhost systemd[1]: libpod-conmon-894eeeb9362ea4a9b32fe78737ccd299953f0e1f11c98181100e46df78d07870.scope: Deactivated successfully. Nov 27 05:07:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:47.279 265123 INFO neutron.agent.dhcp.agent [None req-9b065f0d-7ce9-47a2-ab2f-bdbd69fa1db0 - - - - - -] DHCP configuration for ports {'5b11b185-fbb2-4d32-8799-c379f25a6bbb'} is completed#033[00m Nov 27 05:07:47 localhost nova_compute[284026]: 2025-11-27 10:07:47.522 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:07:47 localhost nova_compute[284026]: 2025-11-27 10:07:47.561 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 05:07:47 localhost nova_compute[284026]: 2025-11-27 10:07:47.562 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:07:47 localhost nova_compute[284026]: 2025-11-27 10:07:47.562 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:07:47 localhost nova_compute[284026]: 2025-11-27 10:07:47.597 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.035s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:07:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e146 e146: 6 total, 6 up, 6 in Nov 27 05:07:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e146 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:48 localhost podman[323558]: Nov 27 05:07:48 localhost podman[323558]: 2025-11-27 10:07:48.04553829 +0000 UTC m=+0.098436321 container create 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:48 localhost systemd[1]: Started libpod-conmon-3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59.scope. Nov 27 05:07:48 localhost podman[323558]: 2025-11-27 10:07:47.997044524 +0000 UTC m=+0.049942595 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:48 localhost systemd[1]: Started libcrun container. Nov 27 05:07:48 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f9f3b5fecc700172406131a1f497d7d0d51e531f297f28e030ad30157726481d/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:48 localhost podman[323558]: 2025-11-27 10:07:48.114214529 +0000 UTC m=+0.167112560 container init 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:07:48 localhost podman[323558]: 2025-11-27 10:07:48.126448049 +0000 UTC m=+0.179346120 container start 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:48 localhost dnsmasq[323576]: started, version 2.85 cachesize 150 Nov 27 05:07:48 localhost dnsmasq[323576]: DNS service limited to local subnets Nov 27 05:07:48 localhost dnsmasq[323576]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:48 localhost dnsmasq[323576]: warning: no upstream servers configured Nov 27 05:07:48 localhost dnsmasq-dhcp[323576]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:48 localhost dnsmasq[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:48 localhost dnsmasq-dhcp[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:48 localhost dnsmasq-dhcp[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:48.437 265123 INFO neutron.agent.dhcp.agent [None req-445de2f9-4ec4-4c11-bbf8-96ff55d6c895 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:48 localhost dnsmasq[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:48 localhost dnsmasq-dhcp[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:48 localhost dnsmasq-dhcp[323576]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:48 localhost podman[323594]: 2025-11-27 10:07:48.571800081 +0000 UTC m=+0.056226595 container kill 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:07:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e147 e147: 6 total, 6 up, 6 in Nov 27 05:07:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:48.889 265123 INFO neutron.agent.dhcp.agent [None req-df3bffbe-a896-4328-b3fb-76d150a73b91 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2047191406' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2047191406' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:49.134 2 INFO neutron.agent.securitygroups_rpc [None req-eb353944-333d-4e85-8982-7f46eeb5c648 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:07:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v304: 177 pgs: 177 active+clean; 889 MiB data, 2.9 GiB used, 39 GiB / 42 GiB avail; 68 KiB/s rd, 37 MiB/s wr, 106 op/s Nov 27 05:07:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:50.045 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '16'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:07:50 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:07:50 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:50 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:50 localhost podman[323630]: 2025-11-27 10:07:50.556851459 +0000 UTC m=+0.067911859 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:50 localhost nova_compute[284026]: 2025-11-27 10:07:50.593 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:50 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:50.925 2 INFO neutron.agent.securitygroups_rpc [None req-82be7730-3e65-4070-8eef-289a25bc8852 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:07:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v305: 177 pgs: 177 active+clean; 1.0 GiB data, 3.4 GiB used, 39 GiB / 42 GiB avail; 125 KiB/s rd, 59 MiB/s wr, 186 op/s Nov 27 05:07:51 localhost nova_compute[284026]: 2025-11-27 10:07:51.248 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:51 localhost dnsmasq[323576]: exiting on receipt of SIGTERM Nov 27 05:07:51 localhost podman[323667]: 2025-11-27 10:07:51.31896479 +0000 UTC m=+0.070899700 container kill 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:51 localhost systemd[1]: libpod-3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59.scope: Deactivated successfully. Nov 27 05:07:51 localhost podman[323679]: 2025-11-27 10:07:51.406137837 +0000 UTC m=+0.073145191 container died 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:07:51 localhost podman[323679]: 2025-11-27 10:07:51.446702379 +0000 UTC m=+0.113709673 container cleanup 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:07:51 localhost systemd[1]: libpod-conmon-3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59.scope: Deactivated successfully. Nov 27 05:07:51 localhost podman[323681]: 2025-11-27 10:07:51.488150526 +0000 UTC m=+0.139342403 container remove 3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:51 localhost systemd[1]: var-lib-containers-storage-overlay-f9f3b5fecc700172406131a1f497d7d0d51e531f297f28e030ad30157726481d-merged.mount: Deactivated successfully. Nov 27 05:07:51 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-3f5a5be9bc2c649324c384d793e1cc162623afec53aae4971621ae9165931d59-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:07:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 8400.1 total, 600.0 interval#012Cumulative writes: 8950 writes, 37K keys, 8950 commit groups, 1.0 writes per commit group, ingest: 0.03 GB, 0.00 MB/s#012Cumulative WAL: 8950 writes, 2239 syncs, 4.00 writes per sync, written: 0.03 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 3818 writes, 14K keys, 3818 commit groups, 1.0 writes per commit group, ingest: 13.21 MB, 0.02 MB/s#012Interval WAL: 3818 writes, 1613 syncs, 2.37 writes per sync, written: 0.01 GB, 0.02 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 05:07:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e148 e148: 6 total, 6 up, 6 in Nov 27 05:07:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e148 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:52 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:52.877 2 INFO neutron.agent.securitygroups_rpc [None req-3119c135-5b6f-4ca4-8389-d453c82a8be9 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v307: 177 pgs: 177 active+clean; 1.0 GiB data, 3.4 GiB used, 39 GiB / 42 GiB avail; 71 KiB/s rd, 29 MiB/s wr, 102 op/s Nov 27 05:07:53 localhost podman[323759]: Nov 27 05:07:53 localhost podman[323759]: 2025-11-27 10:07:53.319016693 +0000 UTC m=+0.098185915 container create 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:07:53 localhost systemd[1]: Started libpod-conmon-80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5.scope. Nov 27 05:07:53 localhost podman[323759]: 2025-11-27 10:07:53.277027692 +0000 UTC m=+0.056196954 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:53 localhost systemd[1]: tmp-crun.1IIPVI.mount: Deactivated successfully. Nov 27 05:07:53 localhost systemd[1]: Started libcrun container. Nov 27 05:07:53 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/aa978e430f81df08138518d4b007f75a0d8dd5d6fc134fea8e91d033073dbef8/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:53 localhost podman[323759]: 2025-11-27 10:07:53.417787023 +0000 UTC m=+0.196956245 container init 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:07:53 localhost podman[323759]: 2025-11-27 10:07:53.427395442 +0000 UTC m=+0.206564664 container start 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:53 localhost dnsmasq[323778]: started, version 2.85 cachesize 150 Nov 27 05:07:53 localhost dnsmasq[323778]: DNS service limited to local subnets Nov 27 05:07:53 localhost dnsmasq[323778]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:53 localhost dnsmasq[323778]: warning: no upstream servers configured Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:07:53 localhost dnsmasq[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:53 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:53.500 265123 INFO neutron.agent.dhcp.agent [None req-08747ec9-2951-4bf7-825d-a0fb854b090e - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:52Z, description=, device_id=, device_owner=, dns_assignment=[, ], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[, ], id=4a4359d4-39de-4bd4-99fd-c612bb827419, ip_allocation=immediate, mac_address=fa:16:3e:d5:77:ce, name=tempest-NetworksTestDHCPv6-1732094002, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:05:19Z, description=, dns_domain=, id=1ba86240-13a7-470b-9c24-36a566ecc5cd, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-NetworksTestDHCPv6-test-network-1513808478, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=41949, qos_policy_id=None, revision_number=61, router:external=False, shared=False, standard_attr_id=1019, status=ACTIVE, subnets=['218106a6-27bd-40f8-a9ac-e7e34d48ede3', '3e64bd94-f268-4a20-b8b8-cd6188a50ae5'], tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:48Z, vlan_transparent=None, network_id=1ba86240-13a7-470b-9c24-36a566ecc5cd, port_security_enabled=True, project_id=96ff25dceebe411dab64c5847327ddc0, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4f0eb495-1507-4658-b9b1-7233f2e8707a'], standard_attr_id=1723, status=DOWN, tags=[], tenant_id=96ff25dceebe411dab64c5847327ddc0, updated_at=2025-11-27T10:07:52Z on network 1ba86240-13a7-470b-9c24-36a566ecc5cd#033[00m Nov 27 05:07:53 localhost dnsmasq[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 2 addresses Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:53 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:53 localhost podman[323798]: 2025-11-27 10:07:53.701117432 +0000 UTC m=+0.062083493 container kill 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:07:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e149 e149: 6 total, 6 up, 6 in Nov 27 05:07:53 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:53.793 265123 INFO neutron.agent.dhcp.agent [None req-a244832a-624c-4731-b55a-cc67cb71f920 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:53 localhost neutron_sriov_agent[258162]: 2025-11-27 10:07:53.914 2 INFO neutron.agent.securitygroups_rpc [None req-32d490d1-720b-4a50-a3eb-10f631df879a 6c3015b24ea642828c6ef69b2911c0cc 96ff25dceebe411dab64c5847327ddc0 - - default default] Security group member updated ['4f0eb495-1507-4658-b9b1-7233f2e8707a']#033[00m Nov 27 05:07:54 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:54.009 265123 INFO neutron.agent.dhcp.agent [None req-bc84e8b6-10c7-4788-ab74-4ee763165e86 - - - - - -] DHCP configuration for ports {'4a4359d4-39de-4bd4-99fd-c612bb827419'} is completed#033[00m Nov 27 05:07:54 localhost dnsmasq[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:54 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:54 localhost dnsmasq-dhcp[323778]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:54 localhost podman[323834]: 2025-11-27 10:07:54.127996676 +0000 UTC m=+0.056302637 container kill 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:07:54 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:54.182 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:07:54Z, description=, device_id=8c55f6ca-b719-4d12-81a5-fba544b1f9d9, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e1e23d10-21bd-4365-99c6-d1ec3e87fbf7, ip_allocation=immediate, mac_address=fa:16:3e:ab:6b:de, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1732, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:07:54Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:07:54 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:07:54 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:07:54 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:07:54 localhost podman[323872]: 2025-11-27 10:07:54.43563308 +0000 UTC m=+0.065122234 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:07:54 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:54.769 265123 INFO neutron.agent.dhcp.agent [None req-b3dd839c-4061-468b-b9c9-1d04bc701597 - - - - - -] DHCP configuration for ports {'e1e23d10-21bd-4365-99c6-d1ec3e87fbf7'} is completed#033[00m Nov 27 05:07:54 localhost nova_compute[284026]: 2025-11-27 10:07:54.848 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v309: 177 pgs: 177 active+clean; 1.0 GiB data, 3.4 GiB used, 39 GiB / 42 GiB avail; 65 KiB/s rd, 27 MiB/s wr, 93 op/s Nov 27 05:07:55 localhost nova_compute[284026]: 2025-11-27 10:07:55.596 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e150 e150: 6 total, 6 up, 6 in Nov 27 05:07:55 localhost openstack_network_exporter[244641]: ERROR 10:07:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:07:55 localhost openstack_network_exporter[244641]: ERROR 10:07:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:07:55 localhost openstack_network_exporter[244641]: ERROR 10:07:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:07:55 localhost openstack_network_exporter[244641]: ERROR 10:07:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:07:55 localhost openstack_network_exporter[244641]: Nov 27 05:07:55 localhost openstack_network_exporter[244641]: ERROR 10:07:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:07:55 localhost openstack_network_exporter[244641]: Nov 27 05:07:55 localhost podman[323908]: 2025-11-27 10:07:55.836380107 +0000 UTC m=+0.076588983 container kill 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:07:55 localhost dnsmasq[323778]: exiting on receipt of SIGTERM Nov 27 05:07:55 localhost systemd[1]: libpod-80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5.scope: Deactivated successfully. Nov 27 05:07:55 localhost podman[323922]: 2025-11-27 10:07:55.905804037 +0000 UTC m=+0.052574177 container died 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:55 localhost podman[323922]: 2025-11-27 10:07:55.947843449 +0000 UTC m=+0.094613519 container cleanup 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:07:55 localhost systemd[1]: libpod-conmon-80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5.scope: Deactivated successfully. Nov 27 05:07:56 localhost podman[323924]: 2025-11-27 10:07:56.013396133 +0000 UTC m=+0.150007189 container remove 80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:07:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 8400.2 total, 600.0 interval#012Cumulative writes: 11K writes, 47K keys, 11K commit groups, 1.0 writes per commit group, ingest: 0.04 GB, 0.00 MB/s#012Cumulative WAL: 11K writes, 3261 syncs, 3.54 writes per sync, written: 0.04 GB, 0.00 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 5761 writes, 21K keys, 5761 commit groups, 1.0 writes per commit group, ingest: 17.82 MB, 0.03 MB/s#012Interval WAL: 5760 writes, 2414 syncs, 2.39 writes per sync, written: 0.02 GB, 0.03 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 05:07:56 localhost nova_compute[284026]: 2025-11-27 10:07:56.286 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:56 localhost systemd[1]: var-lib-containers-storage-overlay-aa978e430f81df08138518d4b007f75a0d8dd5d6fc134fea8e91d033073dbef8-merged.mount: Deactivated successfully. Nov 27 05:07:56 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-80bc4a0604d423fabb5762d5e6870b4d3a2c05298e9f73d580b03fc85ae68cd5-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:57 localhost podman[324002]: Nov 27 05:07:57 localhost podman[324002]: 2025-11-27 10:07:57.013982175 +0000 UTC m=+0.095218595 container create 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:07:57 localhost systemd[1]: Started libpod-conmon-1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab.scope. Nov 27 05:07:57 localhost podman[324002]: 2025-11-27 10:07:56.966171368 +0000 UTC m=+0.047407798 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:07:57 localhost systemd[1]: Started libcrun container. Nov 27 05:07:57 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/e08d905455f61455d29dd0c6af376d2a05a41806c7d8a5b11dc87c4dae1a1f34/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:07:57 localhost podman[324002]: 2025-11-27 10:07:57.09398665 +0000 UTC m=+0.175223070 container init 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:57 localhost podman[324002]: 2025-11-27 10:07:57.102877579 +0000 UTC m=+0.184114009 container start 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:57 localhost dnsmasq[324020]: started, version 2.85 cachesize 150 Nov 27 05:07:57 localhost dnsmasq[324020]: DNS service limited to local subnets Nov 27 05:07:57 localhost dnsmasq[324020]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:07:57 localhost dnsmasq[324020]: warning: no upstream servers configured Nov 27 05:07:57 localhost dnsmasq-dhcp[324020]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:07:57 localhost dnsmasq[324020]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/addn_hosts - 0 addresses Nov 27 05:07:57 localhost dnsmasq-dhcp[324020]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/host Nov 27 05:07:57 localhost dnsmasq-dhcp[324020]: read /var/lib/neutron/dhcp/1ba86240-13a7-470b-9c24-36a566ecc5cd/opts Nov 27 05:07:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v311: 177 pgs: 177 active+clean; 1.1 GiB data, 3.8 GiB used, 38 GiB / 42 GiB avail; 59 KiB/s rd, 17 MiB/s wr, 85 op/s Nov 27 05:07:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:57.324 265123 INFO neutron.agent.dhcp.agent [None req-7202afc7-c4ea-48e5-a170-301b2552dce4 - - - - - -] DHCP configuration for ports {'03eab131-d152-4149-a437-e20cd160cc9e', '2e1b06ae-5268-4e92-89b9-6d951a8114ea'} is completed#033[00m Nov 27 05:07:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:07:57 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2424999366' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:07:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:07:57 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2424999366' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:07:57 localhost dnsmasq[324020]: exiting on receipt of SIGTERM Nov 27 05:07:57 localhost podman[324037]: 2025-11-27 10:07:57.440626143 +0000 UTC m=+0.065435853 container kill 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:07:57 localhost systemd[1]: libpod-1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab.scope: Deactivated successfully. Nov 27 05:07:57 localhost podman[324050]: 2025-11-27 10:07:57.517577645 +0000 UTC m=+0.064756255 container died 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:07:57 localhost podman[324050]: 2025-11-27 10:07:57.547431039 +0000 UTC m=+0.094609589 container cleanup 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:07:57 localhost systemd[1]: libpod-conmon-1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab.scope: Deactivated successfully. Nov 27 05:07:57 localhost podman[324057]: 2025-11-27 10:07:57.593362575 +0000 UTC m=+0.127981807 container remove 1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1ba86240-13a7-470b-9c24-36a566ecc5cd, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:07:57 localhost nova_compute[284026]: 2025-11-27 10:07:57.607 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:57 localhost ovn_controller[156436]: 2025-11-27T10:07:57Z|00246|binding|INFO|Releasing lport 03eab131-d152-4149-a437-e20cd160cc9e from this chassis (sb_readonly=0) Nov 27 05:07:57 localhost kernel: device tap03eab131-d1 left promiscuous mode Nov 27 05:07:57 localhost ovn_controller[156436]: 2025-11-27T10:07:57Z|00247|binding|INFO|Setting lport 03eab131-d152-4149-a437-e20cd160cc9e down in Southbound Nov 27 05:07:57 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:57.617 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:0:1:f816:3eff:fea5:4138/64 2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1ba86240-13a7-470b-9c24-36a566ecc5cd', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '96ff25dceebe411dab64c5847327ddc0', 'neutron:revision_number': '15', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=64e46252-7d38-4a53-9193-71032c8a4f9b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=03eab131-d152-4149-a437-e20cd160cc9e) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:07:57 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:57.619 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 03eab131-d152-4149-a437-e20cd160cc9e in datapath 1ba86240-13a7-470b-9c24-36a566ecc5cd unbound from our chassis#033[00m Nov 27 05:07:57 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:57.621 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 1ba86240-13a7-470b-9c24-36a566ecc5cd, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:07:57 localhost ovn_metadata_agent[162087]: 2025-11-27 10:07:57.622 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[bafa938e-72bc-4043-8cc0-1e36589316cd]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:07:57 localhost nova_compute[284026]: 2025-11-27 10:07:57.637 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:07:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e151 e151: 6 total, 6 up, 6 in Nov 27 05:07:57 localhost systemd[1]: var-lib-containers-storage-overlay-e08d905455f61455d29dd0c6af376d2a05a41806c7d8a5b11dc87c4dae1a1f34-merged.mount: Deactivated successfully. Nov 27 05:07:57 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-1daa0d443468d8f15d333ccb8b09de71ccce9a48402d5ca13094b0928da852ab-userdata-shm.mount: Deactivated successfully. Nov 27 05:07:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e151 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:07:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:07:57.887 265123 INFO neutron.agent.dhcp.agent [None req-18da39ed-755a-4d30-811e-f57c82a17e36 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:07:57 localhost systemd[1]: run-netns-qdhcp\x2d1ba86240\x2d13a7\x2d470b\x2d9c24\x2d36a566ecc5cd.mount: Deactivated successfully. Nov 27 05:07:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v313: 177 pgs: 177 active+clean; 1.1 GiB data, 3.8 GiB used, 38 GiB / 42 GiB avail; 59 KiB/s rd, 17 MiB/s wr, 85 op/s Nov 27 05:08:00 localhost nova_compute[284026]: 2025-11-27 10:08:00.599 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:08:00 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:08:00 localhost podman[324079]: 2025-11-27 10:08:00.990480656 +0000 UTC m=+0.085250137 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 05:08:00 localhost podman[324079]: 2025-11-27 10:08:00.995623205 +0000 UTC m=+0.090392716 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}) Nov 27 05:08:01 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:08:01 localhost systemd[1]: tmp-crun.O3uIAi.mount: Deactivated successfully. Nov 27 05:08:01 localhost podman[324080]: 2025-11-27 10:08:01.040012219 +0000 UTC m=+0.131010168 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125) Nov 27 05:08:01 localhost podman[324080]: 2025-11-27 10:08:01.082106932 +0000 UTC m=+0.173104881 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, container_name=multipathd, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.build-date=20251125) Nov 27 05:08:01 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:08:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v314: 177 pgs: 177 active+clean; 145 MiB data, 844 MiB used, 41 GiB / 42 GiB avail; 111 KiB/s rd, 14 MiB/s wr, 170 op/s Nov 27 05:08:01 localhost nova_compute[284026]: 2025-11-27 10:08:01.321 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:01.521 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:08:01 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:01.988 2 INFO neutron.agent.securitygroups_rpc [None req-6927a1b5-a7b4-4be8-b51a-44a4133cf412 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:08:02 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:08:02 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:08:02 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 6bb09fd3-e23b-4421-b359-19378a96f13b (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:08:02 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 6bb09fd3-e23b-4421-b359-19378a96f13b (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:08:02 localhost ceph-mgr[290377]: [progress INFO root] Completed event 6bb09fd3-e23b-4421-b359-19378a96f13b (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:08:02 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:08:02 localhost ovn_controller[156436]: 2025-11-27T10:08:02Z|00248|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:08:02 localhost nova_compute[284026]: 2025-11-27 10:08:02.218 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e152 e152: 6 total, 6 up, 6 in Nov 27 05:08:02 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:08:02 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:08:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e152 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:03 localhost nova_compute[284026]: 2025-11-27 10:08:03.071 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v316: 177 pgs: 177 active+clean; 145 MiB data, 844 MiB used, 41 GiB / 42 GiB avail; 64 KiB/s rd, 3.2 KiB/s wr, 102 op/s Nov 27 05:08:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:03.480 2 INFO neutron.agent.securitygroups_rpc [None req-aaa46a8a-aabb-48c8-a0e1-f9bfc232bd8d 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:04 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:04.238 2 INFO neutron.agent.securitygroups_rpc [None req-a065c0e2-4803-46df-aabd-f58057716ae0 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e153 e153: 6 total, 6 up, 6 in Nov 27 05:08:04 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:08:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:08:04 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:04.853 2 INFO neutron.agent.securitygroups_rpc [None req-7f6cb490-f280-4fb8-b1d4-237a396abcba 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v318: 177 pgs: 177 active+clean; 145 MiB data, 844 MiB used, 41 GiB / 42 GiB avail; 64 KiB/s rd, 3.2 KiB/s wr, 102 op/s Nov 27 05:08:05 localhost podman[324212]: 2025-11-27 10:08:05.296295275 +0000 UTC m=+0.063088091 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:08:05 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:08:05 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:05 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:05 localhost ovn_controller[156436]: 2025-11-27T10:08:05Z|00249|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:08:05 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:08:05 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:05.483 2 INFO neutron.agent.securitygroups_rpc [None req-756013c8-6c6a-4e2f-bb37-2613a231daa6 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:05 localhost nova_compute[284026]: 2025-11-27 10:08:05.492 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:05 localhost nova_compute[284026]: 2025-11-27 10:08:05.600 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:06 localhost nova_compute[284026]: 2025-11-27 10:08:06.330 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e154 e154: 6 total, 6 up, 6 in Nov 27 05:08:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v320: 177 pgs: 177 active+clean; 145 MiB data, 844 MiB used, 41 GiB / 42 GiB avail; 20 KiB/s rd, 1.7 KiB/s wr, 26 op/s Nov 27 05:08:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e154 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:08 localhost podman[242678]: time="2025-11-27T10:08:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:08:08 localhost podman[242678]: @ - - [27/Nov/2025:10:08:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:08:08 localhost podman[242678]: @ - - [27/Nov/2025:10:08:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19739 "" "Go-http-client/1.1" Nov 27 05:08:08 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:08.876 2 INFO neutron.agent.securitygroups_rpc [None req-ed81919f-6f52-4381-8873-e57e972d1252 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:08:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:08:09 localhost podman[324233]: 2025-11-27 10:08:09.011390476 +0000 UTC m=+0.099028647 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:08:09 localhost podman[324234]: 2025-11-27 10:08:09.070235921 +0000 UTC m=+0.151949692 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:08:09 localhost podman[324233]: 2025-11-27 10:08:09.080226829 +0000 UTC m=+0.167865040 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:08:09 localhost podman[324234]: 2025-11-27 10:08:09.083053855 +0000 UTC m=+0.164767666 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:08:09 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:08:09 localhost systemd[1]: tmp-crun.tZVPLl.mount: Deactivated successfully. Nov 27 05:08:09 localhost podman[324235]: 2025-11-27 10:08:09.166389509 +0000 UTC m=+0.244930405 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.component=ubi9-minimal-container, config_id=edpm, io.buildah.version=1.33.7, build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, architecture=x86_64, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vendor=Red Hat, Inc., vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., container_name=openstack_network_exporter, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, release=1755695350, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 05:08:09 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:08:09 localhost podman[324235]: 2025-11-27 10:08:09.208070572 +0000 UTC m=+0.286611438 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, build-date=2025-08-20T13:12:41, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., release=1755695350, architecture=x86_64, vcs-type=git, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, distribution-scope=public, io.openshift.expose-services=, name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 05:08:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v321: 177 pgs: 177 active+clean; 145 MiB data, 844 MiB used, 41 GiB / 42 GiB avail; 17 KiB/s rd, 1.5 KiB/s wr, 23 op/s Nov 27 05:08:09 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:08:10 localhost nova_compute[284026]: 2025-11-27 10:08:10.602 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:10.677 2 INFO neutron.agent.securitygroups_rpc [None req-4d894b9f-e702-44f8-b2ee-354182461593 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:10.754 2 INFO neutron.agent.securitygroups_rpc [None req-300b19d7-80b5-4980-8166-15bd16654f03 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:11 localhost systemd[1]: tmp-crun.ZJJ11V.mount: Deactivated successfully. Nov 27 05:08:11 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:08:11 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:11 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:11 localhost podman[324309]: 2025-11-27 10:08:11.078470834 +0000 UTC m=+0.069965104 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:08:11 localhost ovn_controller[156436]: 2025-11-27T10:08:11Z|00250|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:08:11 localhost nova_compute[284026]: 2025-11-27 10:08:11.208 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v322: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 2.1 KiB/s wr, 45 op/s Nov 27 05:08:11 localhost nova_compute[284026]: 2025-11-27 10:08:11.332 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:08:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 600.0 total, 600.0 interval#012Cumulative writes: 2129 writes, 22K keys, 2129 commit groups, 1.0 writes per commit group, ingest: 0.04 GB, 0.07 MB/s#012Cumulative WAL: 2129 writes, 2129 syncs, 1.00 writes per sync, written: 0.04 GB, 0.07 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 2129 writes, 22K keys, 2129 commit groups, 1.0 writes per commit group, ingest: 41.51 MB, 0.07 MB/s#012Interval WAL: 2129 writes, 2129 syncs, 1.00 writes per sync, written: 0.04 GB, 0.07 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 165.1 0.17 0.07 7 0.024 0 0 0.0 0.0#012 L6 1/0 17.29 MB 0.0 0.1 0.0 0.1 0.1 0.0 0.0 4.1 138.3 126.5 0.89 0.31 6 0.149 75K 2896 0.0 0.0#012 Sum 1/0 17.29 MB 0.0 0.1 0.0 0.1 0.1 0.0 0.0 5.1 116.3 132.6 1.06 0.37 13 0.082 75K 2896 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.1 0.0 0.1 0.1 0.0 0.0 5.1 116.6 132.9 1.06 0.37 12 0.088 75K 2896 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Low 0/0 0.00 KB 0.0 0.1 0.0 0.1 0.1 0.0 0.0 0.0 138.3 126.5 0.89 0.31 6 0.149 75K 2896 0.0 0.0#012High 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 167.6 0.17 0.07 6 0.028 0 0 0.0 0.0#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 600.0 total, 600.0 interval#012Flush(GB): cumulative 0.027, interval 0.027#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.14 GB write, 0.23 MB/s write, 0.12 GB read, 0.21 MB/s read, 1.1 seconds#012Interval compaction: 0.14 GB write, 0.23 MB/s write, 0.12 GB read, 0.21 MB/s read, 1.1 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x55fa1a6d5350#2 capacity: 308.00 MB usage: 12.26 MB table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 0 last_secs: 8.3e-05 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(542,11.74 MB,3.81045%) FilterBlock(13,233.86 KB,0.0741488%) IndexBlock(13,298.55 KB,0.094659%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] ** Nov 27 05:08:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e155 e155: 6 total, 6 up, 6 in Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #31. Immutable memtables: 0. Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.413028) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 15] Flushing memtable with next log file: 31 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092413067, "job": 15, "event": "flush_started", "num_memtables": 1, "num_entries": 2372, "num_deletes": 265, "total_data_size": 3361829, "memory_usage": 3425152, "flush_reason": "Manual Compaction"} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 15] Level-0 flush table #32: started Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092429387, "cf_name": "default", "job": 15, "event": "table_file_creation", "file_number": 32, "file_size": 2148264, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 20209, "largest_seqno": 22576, "table_properties": {"data_size": 2139456, "index_size": 5377, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2373, "raw_key_size": 19213, "raw_average_key_size": 20, "raw_value_size": 2121355, "raw_average_value_size": 2300, "num_data_blocks": 234, "num_entries": 922, "num_filter_entries": 922, "num_deletions": 265, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237948, "oldest_key_time": 1764237948, "file_creation_time": 1764238092, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 32, "seqno_to_time_mapping": "N/A"}} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 15] Flush lasted 16419 microseconds, and 6213 cpu microseconds. Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.429443) [db/flush_job.cc:967] [default] [JOB 15] Level-0 flush table #32: 2148264 bytes OK Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.429468) [db/memtable_list.cc:519] [default] Level-0 commit table #32 started Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.431975) [db/memtable_list.cc:722] [default] Level-0 commit table #32: memtable #1 done Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.431999) EVENT_LOG_v1 {"time_micros": 1764238092431992, "job": 15, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.432021) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 15] Try to delete WAL files size 3351092, prev total WAL file size 3351092, number of live WAL files 2. Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000028.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.432991) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6C6F676D0034303139' seq:72057594037927935, type:22 .. '6C6F676D0034323730' seq:0, type:0; will stop at (end) Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 16] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 15 Base level 0, inputs: [32(2097KB)], [30(17MB)] Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092433073, "job": 16, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [32], "files_L6": [30], "score": -1, "input_data_size": 20283152, "oldest_snapshot_seqno": -1} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 16] Generated table #33: 12842 keys, 19836365 bytes, temperature: kUnknown Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092577898, "cf_name": "default", "job": 16, "event": "table_file_creation", "file_number": 33, "file_size": 19836365, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19760238, "index_size": 42938, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 32133, "raw_key_size": 343200, "raw_average_key_size": 26, "raw_value_size": 19538747, "raw_average_value_size": 1521, "num_data_blocks": 1639, "num_entries": 12842, "num_filter_entries": 12842, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238092, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 33, "seqno_to_time_mapping": "N/A"}} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.578245) [db/compaction/compaction_job.cc:1663] [default] [JOB 16] Compacted 1@0 + 1@6 files to L6 => 19836365 bytes Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.580446) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 139.9 rd, 136.9 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(2.0, 17.3 +0.0 blob) out(18.9 +0.0 blob), read-write-amplify(18.7) write-amplify(9.2) OK, records in: 13384, records dropped: 542 output_compression: NoCompression Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.580493) EVENT_LOG_v1 {"time_micros": 1764238092580470, "job": 16, "event": "compaction_finished", "compaction_time_micros": 144943, "compaction_time_cpu_micros": 49648, "output_level": 6, "num_output_files": 1, "total_output_size": 19836365, "num_input_records": 13384, "num_output_records": 12842, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000032.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092581135, "job": 16, "event": "table_file_deletion", "file_number": 32} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000030.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238092584886, "job": 16, "event": "table_file_deletion", "file_number": 30} Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.432842) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.584958) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.584965) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.584968) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.584971) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:12.584974) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e155 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:12 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:12.887 2 INFO neutron.agent.securitygroups_rpc [None req-1cd08843-4e6d-43f3-9272-a3826c4c5f3c 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v324: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 2.1 KiB/s wr, 45 op/s Nov 27 05:08:13 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:13.417 2 INFO neutron.agent.securitygroups_rpc [None req-c412dda1-820f-4eb7-8b49-1188dff7b987 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:13 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:13.466 2 INFO neutron.agent.securitygroups_rpc [None req-00fce6ca-b753-4ef8-9707-1b1d1f9d4e5e 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:14 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:14.445 2 INFO neutron.agent.securitygroups_rpc [None req-a079820f-8d90-40a5-a086-d1ec97c914a0 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:15.107 2 INFO neutron.agent.securitygroups_rpc [None req-30491f19-72ab-4ea0-b450-22b80ab37e4b fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v325: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 17 KiB/s rd, 820 B/s wr, 23 op/s Nov 27 05:08:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:15.505 2 INFO neutron.agent.securitygroups_rpc [None req-f3cafc98-29cb-4121-9c19-b7c6a2bbeb6e 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:15 localhost nova_compute[284026]: 2025-11-27 10:08:15.605 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:16.233 2 INFO neutron.agent.securitygroups_rpc [None req-b333d6b4-6254-43f0-a285-517495d5224d 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:16 localhost nova_compute[284026]: 2025-11-27 10:08:16.369 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:16.559 2 INFO neutron.agent.securitygroups_rpc [None req-7911c100-a215-4e3e-9e09-fd16ab67e09e 972bc36852cf4a60bdbf468a78dc139d 4e8f6adffc384df5b92b05bc3eedb3a5 - - default default] Security group member updated ['bde6f4eb-2c7e-4b8d-9001-226551835aaf']#033[00m Nov 27 05:08:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:08:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:08:16 localhost systemd[1]: tmp-crun.3zbZTx.mount: Deactivated successfully. Nov 27 05:08:16 localhost podman[324332]: 2025-11-27 10:08:16.995329693 +0000 UTC m=+0.088678320 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:08:17 localhost podman[324333]: 2025-11-27 10:08:17.044808465 +0000 UTC m=+0.136049495 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:08:17 localhost podman[324332]: 2025-11-27 10:08:17.059898421 +0000 UTC m=+0.153247028 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_controller, io.buildah.version=1.41.3) Nov 27 05:08:17 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:08:17 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:17.075 2 INFO neutron.agent.securitygroups_rpc [None req-2d16737f-ea83-4bd5-b312-afad86129574 de528fab395646f688d426c4ae76ab84 e1afef38cd484609ab77b401f4fdb6aa - - default default] Security group member updated ['1f75e79a-8fb4-4411-ba66-a1adf6582599']#033[00m Nov 27 05:08:17 localhost podman[324333]: 2025-11-27 10:08:17.083950149 +0000 UTC m=+0.175191199 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:08:17 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:08:17 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:17.183 2 INFO neutron.agent.securitygroups_rpc [None req-2d16737f-ea83-4bd5-b312-afad86129574 de528fab395646f688d426c4ae76ab84 e1afef38cd484609ab77b401f4fdb6aa - - default default] Security group member updated ['1f75e79a-8fb4-4411-ba66-a1adf6582599']#033[00m Nov 27 05:08:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v326: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 15 KiB/s rd, 716 B/s wr, 20 op/s Nov 27 05:08:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e155 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:17 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:17.891 2 INFO neutron.agent.securitygroups_rpc [None req-890b27af-db47-4bba-83c2-808f51f230a8 de528fab395646f688d426c4ae76ab84 e1afef38cd484609ab77b401f4fdb6aa - - default default] Security group member updated ['1f75e79a-8fb4-4411-ba66-a1adf6582599']#033[00m Nov 27 05:08:18 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:18.394 2 INFO neutron.agent.securitygroups_rpc [None req-aa8d16dc-d32b-417e-bb50-7d774361c755 de528fab395646f688d426c4ae76ab84 e1afef38cd484609ab77b401f4fdb6aa - - default default] Security group member updated ['1f75e79a-8fb4-4411-ba66-a1adf6582599']#033[00m Nov 27 05:08:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:18.455 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:08:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v327: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 15 KiB/s rd, 716 B/s wr, 20 op/s Nov 27 05:08:20 localhost nova_compute[284026]: 2025-11-27 10:08:20.608 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:20 localhost nova_compute[284026]: 2025-11-27 10:08:20.772 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v328: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 1023 B/s rd, 818 B/s wr, 2 op/s Nov 27 05:08:21 localhost nova_compute[284026]: 2025-11-27 10:08:21.372 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:21 localhost nova_compute[284026]: 2025-11-27 10:08:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e155 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:23.052 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:08:22Z, description=, device_id=8d94bfbf-17c0-45d5-8db7-0091566fa7d8, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=d246be99-618c-4052-8ea8-a70a3bc42507, ip_allocation=immediate, mac_address=fa:16:3e:2f:19:44, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1898, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:08:22Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:08:23 localhost nova_compute[284026]: 2025-11-27 10:08:23.122 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:23.122 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=17, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=16) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:23 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:23.124 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 9 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:08:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v329: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 946 B/s rd, 756 B/s wr, 2 op/s Nov 27 05:08:23 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:08:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:23 localhost podman[324394]: 2025-11-27 10:08:23.283323164 +0000 UTC m=+0.070161141 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:08:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:23.557 265123 INFO neutron.agent.dhcp.agent [None req-de69558a-6509-417b-b107-e4fe3322f83a - - - - - -] DHCP configuration for ports {'d246be99-618c-4052-8ea8-a70a3bc42507'} is completed#033[00m Nov 27 05:08:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:08:23 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/33217197' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:08:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:08:23 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/33217197' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:08:23 localhost nova_compute[284026]: 2025-11-27 10:08:23.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:23 localhost nova_compute[284026]: 2025-11-27 10:08:23.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:24 localhost nova_compute[284026]: 2025-11-27 10:08:24.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:24 localhost nova_compute[284026]: 2025-11-27 10:08:24.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:08:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:25.187 2 INFO neutron.agent.securitygroups_rpc [None req-3002c371-4fc6-41ad-9180-46be240fcee2 c9701462c9a54414ab9401f83e75631f 5c31be19cc454ec3b73720fbdedf3597 - - default default] Security group member updated ['1a1e0009-8374-4d24-a4c2-bb5996b890da']#033[00m Nov 27 05:08:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e156 e156: 6 total, 6 up, 6 in Nov 27 05:08:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v331: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 1023 B/s rd, 818 B/s wr, 2 op/s Nov 27 05:08:25 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:25.449 265123 INFO neutron.agent.linux.ip_lib [None req-03db0276-13ce-4249-af8a-31766565bf97 - - - - - -] Device tapf9a27005-22 cannot be used as it has no MAC address#033[00m Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.474 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost kernel: device tapf9a27005-22 entered promiscuous mode Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.483 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost ovn_controller[156436]: 2025-11-27T10:08:25Z|00251|binding|INFO|Claiming lport f9a27005-2269-493c-8cd9-cdc41db525fc for this chassis. Nov 27 05:08:25 localhost ovn_controller[156436]: 2025-11-27T10:08:25Z|00252|binding|INFO|f9a27005-2269-493c-8cd9-cdc41db525fc: Claiming unknown Nov 27 05:08:25 localhost NetworkManager[5971]: [1764238105.4886] manager: (tapf9a27005-22): new Generic device (/org/freedesktop/NetworkManager/Devices/44) Nov 27 05:08:25 localhost systemd-udevd[324426]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:08:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:25.499 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:0:1::1/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'e1afef38cd484609ab77b401f4fdb6aa', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c5074d92-c511-489f-a76c-04059820fc05, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=f9a27005-2269-493c-8cd9-cdc41db525fc) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:25.500 162092 INFO neutron.agent.ovn.metadata.agent [-] Port f9a27005-2269-493c-8cd9-cdc41db525fc in datapath 1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8 bound to our chassis#033[00m Nov 27 05:08:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:25.502 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:08:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:25.503 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[831113da-2d2e-4818-9762-98578544e619]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost ovn_controller[156436]: 2025-11-27T10:08:25Z|00253|binding|INFO|Setting lport f9a27005-2269-493c-8cd9-cdc41db525fc ovn-installed in OVS Nov 27 05:08:25 localhost ovn_controller[156436]: 2025-11-27T10:08:25Z|00254|binding|INFO|Setting lport f9a27005-2269-493c-8cd9-cdc41db525fc up in Southbound Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.528 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost journal[232028]: ethtool ioctl error on tapf9a27005-22: No such device Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.568 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.598 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost nova_compute[284026]: 2025-11-27 10:08:25.610 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:25 localhost openstack_network_exporter[244641]: ERROR 10:08:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:08:25 localhost openstack_network_exporter[244641]: ERROR 10:08:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:08:25 localhost openstack_network_exporter[244641]: ERROR 10:08:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:08:25 localhost openstack_network_exporter[244641]: ERROR 10:08:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:08:25 localhost openstack_network_exporter[244641]: Nov 27 05:08:25 localhost openstack_network_exporter[244641]: ERROR 10:08:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:08:25 localhost openstack_network_exporter[244641]: Nov 27 05:08:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e157 e157: 6 total, 6 up, 6 in Nov 27 05:08:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:26.247 2 INFO neutron.agent.securitygroups_rpc [None req-74f77aa4-a535-4449-afea-6c03dad23159 c9701462c9a54414ab9401f83e75631f 5c31be19cc454ec3b73720fbdedf3597 - - default default] Security group member updated ['1a1e0009-8374-4d24-a4c2-bb5996b890da']#033[00m Nov 27 05:08:26 localhost nova_compute[284026]: 2025-11-27 10:08:26.396 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:26 localhost podman[324497]: Nov 27 05:08:26 localhost podman[324497]: 2025-11-27 10:08:26.479971587 +0000 UTC m=+0.119890020 container create d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:08:26 localhost systemd[1]: Started libpod-conmon-d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b.scope. Nov 27 05:08:26 localhost podman[324497]: 2025-11-27 10:08:26.428571602 +0000 UTC m=+0.068490035 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:08:26 localhost ovn_controller[156436]: 2025-11-27T10:08:26Z|00255|binding|INFO|Removing iface tapf9a27005-22 ovn-installed in OVS Nov 27 05:08:26 localhost ovn_controller[156436]: 2025-11-27T10:08:26Z|00256|binding|INFO|Removing lport f9a27005-2269-493c-8cd9-cdc41db525fc ovn-installed in OVS Nov 27 05:08:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:26.532 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 2c92bd9a-9b10-428f-924a-4a515dbf78fd with type ""#033[00m Nov 27 05:08:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:26.534 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'e1afef38cd484609ab77b401f4fdb6aa', 'neutron:revision_number': '4', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c5074d92-c511-489f-a76c-04059820fc05, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=f9a27005-2269-493c-8cd9-cdc41db525fc) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:26 localhost nova_compute[284026]: 2025-11-27 10:08:26.535 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:26.537 162092 INFO neutron.agent.ovn.metadata.agent [-] Port f9a27005-2269-493c-8cd9-cdc41db525fc in datapath 1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8 unbound from our chassis#033[00m Nov 27 05:08:26 localhost nova_compute[284026]: 2025-11-27 10:08:26.537 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:26.538 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:08:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:26.539 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[79b9770f-3b61-4ebc-964f-4d74c6b8a57e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:08:26 localhost systemd[1]: tmp-crun.5CAsic.mount: Deactivated successfully. Nov 27 05:08:26 localhost systemd[1]: Started libcrun container. Nov 27 05:08:26 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/f39522a15b97b047962bd6953c5e85b9515326ee895c88fd90458c8a93528a0b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:08:26 localhost podman[324497]: 2025-11-27 10:08:26.582091596 +0000 UTC m=+0.222010029 container init d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:08:26 localhost podman[324497]: 2025-11-27 10:08:26.592851376 +0000 UTC m=+0.232769799 container start d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:08:26 localhost dnsmasq[324515]: started, version 2.85 cachesize 150 Nov 27 05:08:26 localhost dnsmasq[324515]: DNS service limited to local subnets Nov 27 05:08:26 localhost dnsmasq[324515]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:08:26 localhost dnsmasq[324515]: warning: no upstream servers configured Nov 27 05:08:26 localhost dnsmasq-dhcp[324515]: DHCPv6, static leases only on 2001:db8:0:1::, lease time 1d Nov 27 05:08:26 localhost dnsmasq[324515]: read /var/lib/neutron/dhcp/1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8/addn_hosts - 0 addresses Nov 27 05:08:26 localhost dnsmasq-dhcp[324515]: read /var/lib/neutron/dhcp/1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8/host Nov 27 05:08:26 localhost dnsmasq-dhcp[324515]: read /var/lib/neutron/dhcp/1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8/opts Nov 27 05:08:27 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:26.998 265123 INFO neutron.agent.dhcp.agent [None req-01095429-59ae-4295-abac-3b859f530c35 - - - - - -] DHCP configuration for ports {'f59ea60f-65a3-4fe1-908c-6041f532e9db'} is completed#033[00m Nov 27 05:08:27 localhost dnsmasq[324515]: exiting on receipt of SIGTERM Nov 27 05:08:27 localhost podman[324533]: 2025-11-27 10:08:27.156247036 +0000 UTC m=+0.060548232 container kill d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:08:27 localhost systemd[1]: libpod-d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b.scope: Deactivated successfully. Nov 27 05:08:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v333: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 40 KiB/s rd, 3.2 KiB/s wr, 56 op/s Nov 27 05:08:27 localhost podman[324546]: 2025-11-27 10:08:27.232342765 +0000 UTC m=+0.058524507 container died d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, tcib_managed=true, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:08:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e158 e158: 6 total, 6 up, 6 in Nov 27 05:08:27 localhost podman[324546]: 2025-11-27 10:08:27.264219974 +0000 UTC m=+0.090401666 container cleanup d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:08:27 localhost systemd[1]: libpod-conmon-d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b.scope: Deactivated successfully. Nov 27 05:08:27 localhost ovn_controller[156436]: 2025-11-27T10:08:27Z|00257|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:08:27 localhost nova_compute[284026]: 2025-11-27 10:08:27.306 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:27 localhost podman[324548]: 2025-11-27 10:08:27.324069075 +0000 UTC m=+0.143078154 container remove d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-1d0b4930-8e9d-4d2f-aba7-8e9d01e5aff8, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:08:27 localhost nova_compute[284026]: 2025-11-27 10:08:27.338 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:27 localhost kernel: device tapf9a27005-22 left promiscuous mode Nov 27 05:08:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:27.339 2 INFO neutron.agent.securitygroups_rpc [None req-6d79708c-a909-44ab-8888-83851f7c8b0a c9701462c9a54414ab9401f83e75631f 5c31be19cc454ec3b73720fbdedf3597 - - default default] Security group member updated ['1a1e0009-8374-4d24-a4c2-bb5996b890da']#033[00m Nov 27 05:08:27 localhost nova_compute[284026]: 2025-11-27 10:08:27.350 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:27 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:27.377 265123 INFO neutron.agent.dhcp.agent [None req-9bf1c688-33da-4618-8102-190972d22d9e - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:08:27 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:27.378 265123 INFO neutron.agent.dhcp.agent [None req-9bf1c688-33da-4618-8102-190972d22d9e - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:08:27 localhost systemd[1]: var-lib-containers-storage-overlay-f39522a15b97b047962bd6953c5e85b9515326ee895c88fd90458c8a93528a0b-merged.mount: Deactivated successfully. Nov 27 05:08:27 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-d636e892ee12a671f026ff0701845d7ef9adc518c73651730cb1201d157e622b-userdata-shm.mount: Deactivated successfully. Nov 27 05:08:27 localhost systemd[1]: run-netns-qdhcp\x2d1d0b4930\x2d8e9d\x2d4d2f\x2daba7\x2d8e9d01e5aff8.mount: Deactivated successfully. Nov 27 05:08:27 localhost nova_compute[284026]: 2025-11-27 10:08:27.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e158 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:29 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:29.078 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:08:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:29.139 2 INFO neutron.agent.securitygroups_rpc [None req-ce69a6cf-35d6-4b09-b291-f28008cd4957 c9701462c9a54414ab9401f83e75631f 5c31be19cc454ec3b73720fbdedf3597 - - default default] Security group member updated ['1a1e0009-8374-4d24-a4c2-bb5996b890da']#033[00m Nov 27 05:08:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v335: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 51 KiB/s rd, 3.0 KiB/s wr, 70 op/s Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #34. Immutable memtables: 0. Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.289700) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 17] Flushing memtable with next log file: 34 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109289765, "job": 17, "event": "flush_started", "num_memtables": 1, "num_entries": 486, "num_deletes": 251, "total_data_size": 390980, "memory_usage": 400000, "flush_reason": "Manual Compaction"} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 17] Level-0 flush table #35: started Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109294850, "cf_name": "default", "job": 17, "event": "table_file_creation", "file_number": 35, "file_size": 252546, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 22581, "largest_seqno": 23062, "table_properties": {"data_size": 249971, "index_size": 622, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 901, "raw_key_size": 6827, "raw_average_key_size": 19, "raw_value_size": 244635, "raw_average_value_size": 715, "num_data_blocks": 27, "num_entries": 342, "num_filter_entries": 342, "num_deletions": 251, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238093, "oldest_key_time": 1764238093, "file_creation_time": 1764238109, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 35, "seqno_to_time_mapping": "N/A"}} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 17] Flush lasted 5211 microseconds, and 1809 cpu microseconds. Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.294910) [db/flush_job.cc:967] [default] [JOB 17] Level-0 flush table #35: 252546 bytes OK Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.294946) [db/memtable_list.cc:519] [default] Level-0 commit table #35 started Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.299069) [db/memtable_list.cc:722] [default] Level-0 commit table #35: memtable #1 done Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.299099) EVENT_LOG_v1 {"time_micros": 1764238109299091, "job": 17, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.299121) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 17] Try to delete WAL files size 387999, prev total WAL file size 387999, number of live WAL files 2. Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000031.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.300317) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003132303438' seq:72057594037927935, type:22 .. '7061786F73003132333030' seq:0, type:0; will stop at (end) Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 18] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 17 Base level 0, inputs: [35(246KB)], [33(18MB)] Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109300382, "job": 18, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [35], "files_L6": [33], "score": -1, "input_data_size": 20088911, "oldest_snapshot_seqno": -1} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 18] Generated table #36: 12665 keys, 18802226 bytes, temperature: kUnknown Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109418723, "cf_name": "default", "job": 18, "event": "table_file_creation", "file_number": 36, "file_size": 18802226, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 18728317, "index_size": 41168, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 31685, "raw_key_size": 340051, "raw_average_key_size": 26, "raw_value_size": 18510985, "raw_average_value_size": 1461, "num_data_blocks": 1559, "num_entries": 12665, "num_filter_entries": 12665, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238109, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 36, "seqno_to_time_mapping": "N/A"}} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.419143) [db/compaction/compaction_job.cc:1663] [default] [JOB 18] Compacted 1@0 + 1@6 files to L6 => 18802226 bytes Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.422761) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 169.6 rd, 158.7 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.2, 18.9 +0.0 blob) out(17.9 +0.0 blob), read-write-amplify(154.0) write-amplify(74.5) OK, records in: 13184, records dropped: 519 output_compression: NoCompression Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.422792) EVENT_LOG_v1 {"time_micros": 1764238109422779, "job": 18, "event": "compaction_finished", "compaction_time_micros": 118474, "compaction_time_cpu_micros": 56879, "output_level": 6, "num_output_files": 1, "total_output_size": 18802226, "num_input_records": 13184, "num_output_records": 12665, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000035.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109422983, "job": 18, "event": "table_file_deletion", "file_number": 35} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000033.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238109425749, "job": 18, "event": "table_file_deletion", "file_number": 33} Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.299929) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.425879) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.425886) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.425889) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.425895) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:08:29.425897) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:08:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:29.906 2 INFO neutron.agent.securitygroups_rpc [None req-bc816ceb-2115-4476-afdc-4a12ad4d74d0 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:30 localhost nova_compute[284026]: 2025-11-27 10:08:30.613 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:30 localhost nova_compute[284026]: 2025-11-27 10:08:30.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v336: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 134 KiB/s rd, 7.3 KiB/s wr, 182 op/s Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.423 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.752 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.752 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:08:31 localhost nova_compute[284026]: 2025-11-27 10:08:31.752 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:08:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:08:31 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:08:31 localhost systemd[1]: tmp-crun.xkvavV.mount: Deactivated successfully. Nov 27 05:08:31 localhost podman[324586]: 2025-11-27 10:08:31.992630841 +0000 UTC m=+0.083541819 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, container_name=multipathd, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:08:32 localhost podman[324586]: 2025-11-27 10:08:32.006373631 +0000 UTC m=+0.097284599 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=multipathd, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:08:32 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:08:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:08:32 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2376103402' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:08:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:08:32 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2376103402' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:08:32 localhost podman[324585]: 2025-11-27 10:08:32.10212469 +0000 UTC m=+0.193552793 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 05:08:32 localhost podman[324585]: 2025-11-27 10:08:32.106930289 +0000 UTC m=+0.198358352 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:08:32 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:08:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:32.126 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '17'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:08:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:08:32 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3905695956' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.243 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.490s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.304 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.305 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.531 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.533 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11182MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.533 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.534 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:08:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e159 e159: 6 total, 6 up, 6 in Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.697 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.698 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.699 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:08:32 localhost nova_compute[284026]: 2025-11-27 10:08:32.766 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:08:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:32.810 2 INFO neutron.agent.securitygroups_rpc [None req-4e419e3b-1853-4783-8661-9e665ce4cf38 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e159 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v338: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 72 KiB/s rd, 3.7 KiB/s wr, 95 op/s Nov 27 05:08:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:08:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3669703486' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:08:33 localhost nova_compute[284026]: 2025-11-27 10:08:33.253 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.487s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:08:33 localhost nova_compute[284026]: 2025-11-27 10:08:33.261 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:08:33 localhost nova_compute[284026]: 2025-11-27 10:08:33.717 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:08:33 localhost nova_compute[284026]: 2025-11-27 10:08:33.720 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:08:33 localhost nova_compute[284026]: 2025-11-27 10:08:33.720 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.187s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:08:35 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:35.202 2 INFO neutron.agent.securitygroups_rpc [None req-d5310689-4104-407c-a8c7-d646c82b08fd c4a3e2bdfbfc4bc19aee54fe8aa39320 2c1d406c32ca42a9bfe1efc00fc98dd7 - - default default] Security group rule updated ['b7e00b7c-a9e9-479e-84da-a94b8ea964ca']#033[00m Nov 27 05:08:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v339: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 63 KiB/s rd, 3.2 KiB/s wr, 83 op/s Nov 27 05:08:35 localhost nova_compute[284026]: 2025-11-27 10:08:35.617 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:35 localhost nova_compute[284026]: 2025-11-27 10:08:35.721 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:08:35 localhost nova_compute[284026]: 2025-11-27 10:08:35.721 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:08:35 localhost nova_compute[284026]: 2025-11-27 10:08:35.722 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:08:36 localhost nova_compute[284026]: 2025-11-27 10:08:36.065 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:08:36 localhost nova_compute[284026]: 2025-11-27 10:08:36.066 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:08:36 localhost nova_compute[284026]: 2025-11-27 10:08:36.066 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:08:36 localhost nova_compute[284026]: 2025-11-27 10:08:36.067 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:08:36 localhost podman[324667]: 2025-11-27 10:08:36.282480629 +0000 UTC m=+0.063218563 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:08:36 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:08:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:36 localhost nova_compute[284026]: 2025-11-27 10:08:36.425 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:36.649 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:08:36Z, description=, device_id=250a4074-4060-4cc7-b1e9-ef88da3178a0, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=03616dd4-2c2b-478b-99bb-4e5dddaaa469, ip_allocation=immediate, mac_address=fa:16:3e:2f:de:f7, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=1981, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:08:36Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:08:36 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:08:36 localhost podman[324706]: 2025-11-27 10:08:36.831322357 +0000 UTC m=+0.043501792 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:08:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:36 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:37 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:37.124 265123 INFO neutron.agent.dhcp.agent [None req-73eda325-6cd7-495b-856b-f6ed4686a338 - - - - - -] DHCP configuration for ports {'03616dd4-2c2b-478b-99bb-4e5dddaaa469'} is completed#033[00m Nov 27 05:08:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v340: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 2.1 MiB/s rd, 3.2 KiB/s wr, 104 op/s Nov 27 05:08:37 localhost nova_compute[284026]: 2025-11-27 10:08:37.338 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:08:37 localhost nova_compute[284026]: 2025-11-27 10:08:37.372 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:08:37 localhost nova_compute[284026]: 2025-11-27 10:08:37.373 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:08:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e159 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:38 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:38.700 265123 INFO neutron.agent.linux.ip_lib [None req-a8b13e43-fa1e-4201-88c8-def8d74df40b - - - - - -] Device tapc98b1720-02 cannot be used as it has no MAC address#033[00m Nov 27 05:08:38 localhost nova_compute[284026]: 2025-11-27 10:08:38.725 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:38 localhost kernel: device tapc98b1720-02 entered promiscuous mode Nov 27 05:08:38 localhost NetworkManager[5971]: [1764238118.7324] manager: (tapc98b1720-02): new Generic device (/org/freedesktop/NetworkManager/Devices/45) Nov 27 05:08:38 localhost ovn_controller[156436]: 2025-11-27T10:08:38Z|00258|binding|INFO|Claiming lport c98b1720-0211-488c-964d-21d0073c26d1 for this chassis. Nov 27 05:08:38 localhost ovn_controller[156436]: 2025-11-27T10:08:38Z|00259|binding|INFO|c98b1720-0211-488c-964d-21d0073c26d1: Claiming unknown Nov 27 05:08:38 localhost nova_compute[284026]: 2025-11-27 10:08:38.734 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:38 localhost systemd-udevd[324738]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:08:38 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:38.748 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.255.242/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-bd45c14b-d688-4b39-bb9e-8d0539609d09', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-bd45c14b-d688-4b39-bb9e-8d0539609d09', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '04d6be087105405690248faa72ceeb41', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=df5df163-ea1e-4879-b1cb-bdc0a447121e, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=c98b1720-0211-488c-964d-21d0073c26d1) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:38 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:38.750 162092 INFO neutron.agent.ovn.metadata.agent [-] Port c98b1720-0211-488c-964d-21d0073c26d1 in datapath bd45c14b-d688-4b39-bb9e-8d0539609d09 bound to our chassis#033[00m Nov 27 05:08:38 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:38.752 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network bd45c14b-d688-4b39-bb9e-8d0539609d09 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:08:38 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:38.752 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[fa688188-ed14-4dd8-829e-78cde7807acd]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost podman[242678]: time="2025-11-27T10:08:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost ovn_controller[156436]: 2025-11-27T10:08:38Z|00260|binding|INFO|Setting lport c98b1720-0211-488c-964d-21d0073c26d1 ovn-installed in OVS Nov 27 05:08:38 localhost ovn_controller[156436]: 2025-11-27T10:08:38Z|00261|binding|INFO|Setting lport c98b1720-0211-488c-964d-21d0073c26d1 up in Southbound Nov 27 05:08:38 localhost nova_compute[284026]: 2025-11-27 10:08:38.818 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:38 localhost podman[242678]: @ - - [27/Nov/2025:10:08:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost journal[232028]: ethtool ioctl error on tapc98b1720-02: No such device Nov 27 05:08:38 localhost nova_compute[284026]: 2025-11-27 10:08:38.845 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:38 localhost podman[242678]: @ - - [27/Nov/2025:10:08:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19736 "" "Go-http-client/1.1" Nov 27 05:08:38 localhost nova_compute[284026]: 2025-11-27 10:08:38.869 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v341: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 2.1 MiB/s rd, 3.2 KiB/s wr, 103 op/s Nov 27 05:08:39 localhost podman[324811]: Nov 27 05:08:39 localhost podman[324811]: 2025-11-27 10:08:39.680032483 +0000 UTC m=+0.090804377 container create 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:08:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:08:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:08:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:08:39 localhost systemd[1]: Started libpod-conmon-39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e.scope. Nov 27 05:08:39 localhost systemd[1]: tmp-crun.ZLcuu2.mount: Deactivated successfully. Nov 27 05:08:39 localhost podman[324811]: 2025-11-27 10:08:39.637827236 +0000 UTC m=+0.048599160 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:08:39 localhost systemd[1]: Started libcrun container. Nov 27 05:08:39 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/02d8eedafbedc7999cbf7efe0c4caaa867535d7079c0a9ed9f3d7cb5fdc2db1b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:08:39 localhost podman[324811]: 2025-11-27 10:08:39.757550779 +0000 UTC m=+0.168322653 container init 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:08:39 localhost dnsmasq[324862]: started, version 2.85 cachesize 150 Nov 27 05:08:39 localhost dnsmasq[324862]: DNS service limited to local subnets Nov 27 05:08:39 localhost dnsmasq[324862]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:08:39 localhost dnsmasq[324862]: warning: no upstream servers configured Nov 27 05:08:39 localhost dnsmasq-dhcp[324862]: DHCP, static leases only on 10.100.255.240, lease time 1d Nov 27 05:08:39 localhost dnsmasq[324862]: read /var/lib/neutron/dhcp/bd45c14b-d688-4b39-bb9e-8d0539609d09/addn_hosts - 0 addresses Nov 27 05:08:39 localhost dnsmasq-dhcp[324862]: read /var/lib/neutron/dhcp/bd45c14b-d688-4b39-bb9e-8d0539609d09/host Nov 27 05:08:39 localhost dnsmasq-dhcp[324862]: read /var/lib/neutron/dhcp/bd45c14b-d688-4b39-bb9e-8d0539609d09/opts Nov 27 05:08:39 localhost podman[324825]: 2025-11-27 10:08:39.792261244 +0000 UTC m=+0.078848075 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:08:39 localhost podman[324825]: 2025-11-27 10:08:39.798623685 +0000 UTC m=+0.085210726 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:08:39 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:08:39 localhost podman[324826]: 2025-11-27 10:08:39.845690893 +0000 UTC m=+0.133630010 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., release=1755695350, managed_by=edpm_ansible, vendor=Red Hat, Inc., io.openshift.expose-services=, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, maintainer=Red Hat, Inc., io.openshift.tags=minimal rhel9, architecture=x86_64, container_name=openstack_network_exporter, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, build-date=2025-08-20T13:12:41) Nov 27 05:08:39 localhost podman[324826]: 2025-11-27 10:08:39.857891161 +0000 UTC m=+0.145830298 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, vendor=Red Hat, Inc., config_id=edpm, io.openshift.expose-services=, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, name=ubi9-minimal, version=9.6) Nov 27 05:08:39 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:08:39 localhost podman[324811]: 2025-11-27 10:08:39.875014652 +0000 UTC m=+0.285786506 container start 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:08:39 localhost podman[324824]: 2025-11-27 10:08:39.910245891 +0000 UTC m=+0.201376653 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_managed=true, config_id=edpm, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute) Nov 27 05:08:39 localhost podman[324824]: 2025-11-27 10:08:39.920741274 +0000 UTC m=+0.211872036 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:08:39 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:08:40 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:40.110 265123 INFO neutron.agent.dhcp.agent [None req-0299270a-f730-4910-939e-bd2fe789ca8e - - - - - -] DHCP configuration for ports {'3c354697-a93a-4033-9109-0c1f2ef5e396'} is completed#033[00m Nov 27 05:08:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:40.140 2 INFO neutron.agent.securitygroups_rpc [None req-2f3d85d9-11f8-47c1-9953-89e00f4ea2cc fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:40 localhost nova_compute[284026]: 2025-11-27 10:08:40.621 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:40.965 2 INFO neutron.agent.securitygroups_rpc [None req-0da0948e-2a40-4fde-9b74-587029914f45 fc68a00025b24b6a9319a067bd375096 079849d2dcd64e79a824a2adaa804746 - - default default] Security group member updated ['dfc8a325-8ff7-4c07-b23e-ffebc184fe8c']#033[00m Nov 27 05:08:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:08:41 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2759065128' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:08:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v342: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 2.1 MiB/s rd, 1.4 KiB/s wr, 39 op/s Nov 27 05:08:41 localhost nova_compute[284026]: 2025-11-27 10:08:41.469 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e160 e160: 6 total, 6 up, 6 in Nov 27 05:08:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e161 e161: 6 total, 6 up, 6 in Nov 27 05:08:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:42.835 2 INFO neutron.agent.securitygroups_rpc [None req-dd1065e5-7321-4e4c-bea2-3a1caa51494f 28a3eb782d6449649852fe0cd4d88d9e 4100252449c74369989f2c37847377b0 - - default default] Security group member updated ['cabdbaa0-e27a-4c73-b78f-51c6b4725ad5']#033[00m Nov 27 05:08:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e161 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v345: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail; 2.6 MiB/s rd, 1.7 KiB/s wr, 49 op/s Nov 27 05:08:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:43.577 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:08:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:43.577 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:08:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:43.578 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:08:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e162 e162: 6 total, 6 up, 6 in Nov 27 05:08:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:08:44 Nov 27 05:08:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:08:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:08:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['images', 'backups', 'volumes', '.mgr', 'manila_data', 'vms', 'manila_metadata'] Nov 27 05:08:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:08:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:08:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e163 e163: 6 total, 6 up, 6 in Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 2.9989356504745952e-06 of space, bias 1.0, pg target 0.0005987874848780942 quantized to 32 (current 32) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:08:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 2.7263051367950866e-06 of space, bias 4.0, pg target 0.002170138888888889 quantized to 16 (current 16) Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:08:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:08:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v348: 177 pgs: 177 active+clean; 145 MiB data, 845 MiB used, 41 GiB / 42 GiB avail Nov 27 05:08:45 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:08:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:45 localhost podman[324909]: 2025-11-27 10:08:45.354127892 +0000 UTC m=+0.065667438 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:08:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:45 localhost nova_compute[284026]: 2025-11-27 10:08:45.624 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:46 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:46.122+0000 7f54df954640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:46.122+0000 7f54df954640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:46.122+0000 7f54df954640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:46.122+0000 7f54df954640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:46.122+0000 7f54df954640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta' Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "format": "json"}]: dispatch Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:46 localhost nova_compute[284026]: 2025-11-27 10:08:46.474 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v349: 177 pgs: 177 active+clean; 192 MiB data, 909 MiB used, 41 GiB / 42 GiB avail; 73 KiB/s rd, 3.8 MiB/s wr, 111 op/s Nov 27 05:08:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:08:47 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2225804419' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:08:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:08:47 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2225804419' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:08:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:08:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:08:47 localhost systemd[1]: tmp-crun.kayudh.mount: Deactivated successfully. Nov 27 05:08:47 localhost podman[324946]: 2025-11-27 10:08:47.9968617 +0000 UTC m=+0.089230153 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:08:48 localhost podman[324946]: 2025-11-27 10:08:48.033565639 +0000 UTC m=+0.125934092 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:08:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e163 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:48 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:08:48 localhost podman[324947]: 2025-11-27 10:08:48.057291197 +0000 UTC m=+0.146376762 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:08:48 localhost podman[324947]: 2025-11-27 10:08:48.063140185 +0000 UTC m=+0.152225740 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:08:48 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:08:48 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:48.491 2 INFO neutron.agent.securitygroups_rpc [None req-69a63517-9ecb-4c92-9d71-5f4f72f5c2a0 28a3eb782d6449649852fe0cd4d88d9e 4100252449c74369989f2c37847377b0 - - default default] Security group member updated ['cabdbaa0-e27a-4c73-b78f-51c6b4725ad5']#033[00m Nov 27 05:08:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:08:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2386960373' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:08:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v350: 177 pgs: 177 active+clean; 192 MiB data, 909 MiB used, 41 GiB / 42 GiB avail; 62 KiB/s rd, 3.2 MiB/s wr, 94 op/s Nov 27 05:08:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "snap_name": "2a72bac0-d975-4730-9071-4dfa42f94cc1", "format": "json"}]: dispatch Nov 27 05:08:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e164 e164: 6 total, 6 up, 6 in Nov 27 05:08:49 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #44. Immutable memtables: 1. Nov 27 05:08:50 localhost nova_compute[284026]: 2025-11-27 10:08:50.628 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v352: 177 pgs: 177 active+clean; 145 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 111 KiB/s rd, 2.8 MiB/s wr, 162 op/s Nov 27 05:08:51 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:08:51 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:08:51 localhost podman[325010]: 2025-11-27 10:08:51.352953306 +0000 UTC m=+0.051355273 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:08:51 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:08:51 localhost nova_compute[284026]: 2025-11-27 10:08:51.471 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e165 e165: 6 total, 6 up, 6 in Nov 27 05:08:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:08:52 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2104556628' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:08:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:08:52 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2104556628' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:08:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e166 e166: 6 total, 6 up, 6 in Nov 27 05:08:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e166 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v355: 177 pgs: 177 active+clean; 145 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 72 KiB/s rd, 8.8 KiB/s wr, 99 op/s Nov 27 05:08:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "snap_name": "2a72bac0-d975-4730-9071-4dfa42f94cc1_98b7fe29-e49e-4292-a74d-215fd2f0a512", "force": true, "format": "json"}]: dispatch Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1_98b7fe29-e49e-4292-a74d-215fd2f0a512, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta' Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1_98b7fe29-e49e-4292-a74d-215fd2f0a512, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "snap_name": "2a72bac0-d975-4730-9071-4dfa42f94cc1", "force": true, "format": "json"}]: dispatch Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e167 e167: 6 total, 6 up, 6 in Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta.tmp' to config b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5/.meta' Nov 27 05:08:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:2a72bac0-d975-4730-9071-4dfa42f94cc1, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v357: 177 pgs: 177 active+clean; 145 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 80 KiB/s rd, 9.8 KiB/s wr, 111 op/s Nov 27 05:08:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e168 e168: 6 total, 6 up, 6 in Nov 27 05:08:55 localhost nova_compute[284026]: 2025-11-27 10:08:55.630 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:55 localhost openstack_network_exporter[244641]: ERROR 10:08:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:08:55 localhost openstack_network_exporter[244641]: ERROR 10:08:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:08:55 localhost openstack_network_exporter[244641]: ERROR 10:08:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:08:55 localhost openstack_network_exporter[244641]: ERROR 10:08:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:08:55 localhost openstack_network_exporter[244641]: Nov 27 05:08:55 localhost openstack_network_exporter[244641]: ERROR 10:08:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:08:55 localhost openstack_network_exporter[244641]: Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta' Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "format": "json"}]: dispatch Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:56 localhost nova_compute[284026]: 2025-11-27 10:08:56.504 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "format": "json"}]: dispatch Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:08:56 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ebc0337d-35ed-48c3-ba80-7cbaba5a07d5' of type subvolume Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.826+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ebc0337d-35ed-48c3-ba80-7cbaba5a07d5' of type subvolume Nov 27 05:08:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "ebc0337d-35ed-48c3-ba80-7cbaba5a07d5", "force": true, "format": "json"}]: dispatch Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/ebc0337d-35ed-48c3-ba80-7cbaba5a07d5'' moved to trashcan Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:08:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ebc0337d-35ed-48c3-ba80-7cbaba5a07d5, vol_name:cephfs) < "" Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.850+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.850+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.850+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.850+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.850+0000 7f54e1958640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.889+0000 7f54e2159640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.889+0000 7f54e2159640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.889+0000 7f54e2159640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.889+0000 7f54e2159640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:08:56.889+0000 7f54e2159640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:08:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v359: 177 pgs: 177 active+clean; 145 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 76 KiB/s rd, 14 KiB/s wr, 105 op/s Nov 27 05:08:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e169 e169: 6 total, 6 up, 6 in Nov 27 05:08:57 localhost neutron_sriov_agent[258162]: 2025-11-27 10:08:57.787 2 INFO neutron.agent.securitygroups_rpc [None req-d2f2b169-d161-43bc-bd33-9743bbead3df 9b5b3e11fd7b42b09c7853ba4e239182 051c830b422344e3b72d8e2179bed0cf - - default default] Security group member updated ['4fe4d4b3-57c3-499e-81fd-ba692ac93c75']#033[00m Nov 27 05:08:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e169 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:08:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:08:58.855 265123 INFO neutron.agent.linux.ip_lib [None req-f6dfe8f3-e825-4a7b-8859-967abd7fb33c - - - - - -] Device tap562e902d-35 cannot be used as it has no MAC address#033[00m Nov 27 05:08:58 localhost nova_compute[284026]: 2025-11-27 10:08:58.934 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:58 localhost kernel: device tap562e902d-35 entered promiscuous mode Nov 27 05:08:58 localhost NetworkManager[5971]: [1764238138.9494] manager: (tap562e902d-35): new Generic device (/org/freedesktop/NetworkManager/Devices/46) Nov 27 05:08:58 localhost nova_compute[284026]: 2025-11-27 10:08:58.949 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:58 localhost ovn_controller[156436]: 2025-11-27T10:08:58Z|00262|binding|INFO|Claiming lport 562e902d-35f9-4831-825a-fecf549acc87 for this chassis. Nov 27 05:08:58 localhost ovn_controller[156436]: 2025-11-27T10:08:58Z|00263|binding|INFO|562e902d-35f9-4831-825a-fecf549acc87: Claiming unknown Nov 27 05:08:58 localhost systemd-udevd[325065]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:08:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:58.961 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '04d6be087105405690248faa72ceeb41', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=4a9320e8-713e-4120-8b3e-6a4347c6ed0b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=562e902d-35f9-4831-825a-fecf549acc87) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:58.962 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 562e902d-35f9-4831-825a-fecf549acc87 in datapath 49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f bound to our chassis#033[00m Nov 27 05:08:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:58.963 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:08:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:58.964 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[5cf54b4f-3124-4b65-9eb8-0f6e93615d30]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:08:58 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:58 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:58 localhost nova_compute[284026]: 2025-11-27 10:08:58.996 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:58 localhost ovn_controller[156436]: 2025-11-27T10:08:58Z|00264|binding|INFO|Setting lport 562e902d-35f9-4831-825a-fecf549acc87 ovn-installed in OVS Nov 27 05:08:58 localhost ovn_controller[156436]: 2025-11-27T10:08:58Z|00265|binding|INFO|Setting lport 562e902d-35f9-4831-825a-fecf549acc87 up in Southbound Nov 27 05:08:58 localhost nova_compute[284026]: 2025-11-27 10:08:58.999 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost journal[232028]: ethtool ioctl error on tap562e902d-35: No such device Nov 27 05:08:59 localhost nova_compute[284026]: 2025-11-27 10:08:59.045 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:59 localhost nova_compute[284026]: 2025-11-27 10:08:59.078 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "snap_name": "11e98a01-0b96-4bb6-9149-1181ae49d777", "format": "json"}]: dispatch Nov 27 05:08:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:08:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v361: 177 pgs: 177 active+clean; 145 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 67 KiB/s rd, 13 KiB/s wr, 94 op/s Nov 27 05:08:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e170 e170: 6 total, 6 up, 6 in Nov 27 05:08:59 localhost ovn_controller[156436]: 2025-11-27T10:08:59Z|00266|binding|INFO|Removing iface tap562e902d-35 ovn-installed in OVS Nov 27 05:08:59 localhost ovn_controller[156436]: 2025-11-27T10:08:59Z|00267|binding|INFO|Removing lport 562e902d-35f9-4831-825a-fecf549acc87 ovn-installed in OVS Nov 27 05:08:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:59.896 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 1fc1e313-7e1d-4b00-b24c-a555da707c4f with type ""#033[00m Nov 27 05:08:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:59.898 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '04d6be087105405690248faa72ceeb41', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=4a9320e8-713e-4120-8b3e-6a4347c6ed0b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=562e902d-35f9-4831-825a-fecf549acc87) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:08:59 localhost nova_compute[284026]: 2025-11-27 10:08:59.902 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:08:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:59.907 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 562e902d-35f9-4831-825a-fecf549acc87 in datapath 49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f unbound from our chassis#033[00m Nov 27 05:08:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:59.910 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:08:59 localhost ovn_metadata_agent[162087]: 2025-11-27 10:08:59.911 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[08c68e49-1fae-4975-bc9c-7fcbf98e44fa]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:00 localhost podman[325137]: Nov 27 05:09:00 localhost podman[325137]: 2025-11-27 10:09:00.201933144 +0000 UTC m=+0.102132250 container create 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:09:00 localhost podman[325137]: 2025-11-27 10:09:00.156917222 +0000 UTC m=+0.057116328 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:09:00 localhost systemd[1]: Started libpod-conmon-9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf.scope. Nov 27 05:09:00 localhost systemd[1]: tmp-crun.Y5YBpX.mount: Deactivated successfully. Nov 27 05:09:00 localhost systemd[1]: Started libcrun container. Nov 27 05:09:00 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/28a6cffe96054965847c957459d31263e4ba9e4f5095997e7147cdc79b46f43a/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:09:00 localhost podman[325137]: 2025-11-27 10:09:00.321282558 +0000 UTC m=+0.221481674 container init 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:09:00 localhost podman[325137]: 2025-11-27 10:09:00.3314055 +0000 UTC m=+0.231604626 container start 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) Nov 27 05:09:00 localhost dnsmasq[325156]: started, version 2.85 cachesize 150 Nov 27 05:09:00 localhost dnsmasq[325156]: DNS service limited to local subnets Nov 27 05:09:00 localhost dnsmasq[325156]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:09:00 localhost dnsmasq[325156]: warning: no upstream servers configured Nov 27 05:09:00 localhost dnsmasq-dhcp[325156]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:09:00 localhost dnsmasq[325156]: read /var/lib/neutron/dhcp/49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f/addn_hosts - 0 addresses Nov 27 05:09:00 localhost dnsmasq-dhcp[325156]: read /var/lib/neutron/dhcp/49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f/host Nov 27 05:09:00 localhost dnsmasq-dhcp[325156]: read /var/lib/neutron/dhcp/49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f/opts Nov 27 05:09:00 localhost ovn_controller[156436]: 2025-11-27T10:09:00Z|00268|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:09:00 localhost nova_compute[284026]: 2025-11-27 10:09:00.388 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:00.505 265123 INFO neutron.agent.dhcp.agent [None req-d22557c9-9a9e-4a80-9938-d614036ed6cc - - - - - -] DHCP configuration for ports {'75a01298-e4da-4a55-9b32-821e7068a930'} is completed#033[00m Nov 27 05:09:00 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:00.507 2 INFO neutron.agent.securitygroups_rpc [None req-89c451d0-a3fd-42fb-9211-8815006169dd 9b5b3e11fd7b42b09c7853ba4e239182 051c830b422344e3b72d8e2179bed0cf - - default default] Security group member updated ['4fe4d4b3-57c3-499e-81fd-ba692ac93c75']#033[00m Nov 27 05:09:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:00.570 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=18, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=17) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:00 localhost nova_compute[284026]: 2025-11-27 10:09:00.572 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:00.572 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 9 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:09:00 localhost nova_compute[284026]: 2025-11-27 10:09:00.633 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:00 localhost dnsmasq[325156]: exiting on receipt of SIGTERM Nov 27 05:09:00 localhost podman[325173]: 2025-11-27 10:09:00.646969078 +0000 UTC m=+0.072366640 container kill 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:09:00 localhost systemd[1]: libpod-9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf.scope: Deactivated successfully. Nov 27 05:09:00 localhost podman[325187]: 2025-11-27 10:09:00.719602743 +0000 UTC m=+0.060285104 container died 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:09:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e171 e171: 6 total, 6 up, 6 in Nov 27 05:09:00 localhost podman[325187]: 2025-11-27 10:09:00.757962047 +0000 UTC m=+0.098644368 container cleanup 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:09:00 localhost systemd[1]: libpod-conmon-9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf.scope: Deactivated successfully. Nov 27 05:09:00 localhost podman[325194]: 2025-11-27 10:09:00.817019417 +0000 UTC m=+0.143036193 container remove 9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-49c78d0a-1bfd-4cc6-8a11-d0bb4067e24f, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:09:00 localhost nova_compute[284026]: 2025-11-27 10:09:00.831 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:00 localhost kernel: device tap562e902d-35 left promiscuous mode Nov 27 05:09:00 localhost nova_compute[284026]: 2025-11-27 10:09:00.844 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:00.933 265123 INFO neutron.agent.dhcp.agent [None req-32159ea7-751c-4934-8a2a-61075074a52a - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:00.934 265123 INFO neutron.agent.dhcp.agent [None req-32159ea7-751c-4934-8a2a-61075074a52a - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:01 localhost systemd[1]: var-lib-containers-storage-overlay-28a6cffe96054965847c957459d31263e4ba9e4f5095997e7147cdc79b46f43a-merged.mount: Deactivated successfully. Nov 27 05:09:01 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-9f08ee37d6ae3e190574d503935215e91890a7db0d10521adb446f7534253dbf-userdata-shm.mount: Deactivated successfully. Nov 27 05:09:01 localhost systemd[1]: run-netns-qdhcp\x2d49c78d0a\x2d1bfd\x2d4cc6\x2d8a11\x2dd0bb4067e24f.mount: Deactivated successfully. Nov 27 05:09:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v364: 177 pgs: 177 active+clean; 146 MiB data, 894 MiB used, 41 GiB / 42 GiB avail; 61 KiB/s rd, 18 KiB/s wr, 87 op/s Nov 27 05:09:01 localhost nova_compute[284026]: 2025-11-27 10:09:01.536 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:01 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:01.620 2 INFO neutron.agent.securitygroups_rpc [None req-2b61ae3f-8f0f-4cc9-8f83-c067e8d7479d 9a505de2215d44919cd3f3867df24bfb d0796a19572f46d9a1a2de06fab9db68 - - default default] Security group member updated ['ebbb614f-b60c-4f3e-8d62-ebbe70950da7']#033[00m Nov 27 05:09:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:01.681 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:09:01Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=be18cd1b-7c1d-41b4-a52f-b24a90512406, ip_allocation=immediate, mac_address=fa:16:3e:f4:7f:f4, name=tempest-RoutersAdminNegativeTest-1424960965, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=True, project_id=d0796a19572f46d9a1a2de06fab9db68, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['ebbb614f-b60c-4f3e-8d62-ebbe70950da7'], standard_attr_id=2111, status=DOWN, tags=[], tenant_id=d0796a19572f46d9a1a2de06fab9db68, updated_at=2025-11-27T10:09:01Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:09:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e172 e172: 6 total, 6 up, 6 in Nov 27 05:09:01 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:09:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:09:01 localhost podman[325236]: 2025-11-27 10:09:01.961261506 +0000 UTC m=+0.080112737 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3) Nov 27 05:09:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:09:02 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:02.344 265123 INFO neutron.agent.dhcp.agent [None req-09613554-e781-44e5-8f95-1b7b4901d014 - - - - - -] DHCP configuration for ports {'be18cd1b-7c1d-41b4-a52f-b24a90512406'} is completed#033[00m Nov 27 05:09:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:09:02 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:09:02 localhost systemd[1]: tmp-crun.UNmx2N.mount: Deactivated successfully. Nov 27 05:09:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e173 e173: 6 total, 6 up, 6 in Nov 27 05:09:02 localhost podman[325273]: 2025-11-27 10:09:02.583169132 +0000 UTC m=+0.110283050 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:09:02 localhost podman[325273]: 2025-11-27 10:09:02.594143807 +0000 UTC m=+0.121257785 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_managed=true, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:09:02 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:09:02 localhost podman[325275]: 2025-11-27 10:09:02.689220848 +0000 UTC m=+0.215911405 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:09:02 localhost podman[325275]: 2025-11-27 10:09:02.731193558 +0000 UTC m=+0.257884125 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:09:02 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:09:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "snap_name": "11e98a01-0b96-4bb6-9149-1181ae49d777_6d63dfd3-eac9-4abf-bcec-105b4e00baa0", "force": true, "format": "json"}]: dispatch Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777_6d63dfd3-eac9-4abf-bcec-105b4e00baa0, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta' Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777_6d63dfd3-eac9-4abf-bcec-105b4e00baa0, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "snap_name": "11e98a01-0b96-4bb6-9149-1181ae49d777", "force": true, "format": "json"}]: dispatch Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' Nov 27 05:09:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta.tmp' to config b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f/.meta' Nov 27 05:09:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:11e98a01-0b96-4bb6-9149-1181ae49d777, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e173 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Nov 27 05:09:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v367: 177 pgs: 177 active+clean; 146 MiB data, 894 MiB used, 41 GiB / 42 GiB avail; 85 KiB/s rd, 25 KiB/s wr, 122 op/s Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:09:03 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:09:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:09:03 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev fa86c379-0afe-4774-90c2-e7c5452645a0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:09:03 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev fa86c379-0afe-4774-90c2-e7c5452645a0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:09:03 localhost ceph-mgr[290377]: [progress INFO root] Completed event fa86c379-0afe-4774-90c2-e7c5452645a0 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:09:03 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:09:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e174 e174: 6 total, 6 up, 6 in Nov 27 05:09:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:09:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:09:04 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:04.116 2 INFO neutron.agent.securitygroups_rpc [None req-230bb7b6-68cf-4c4a-954c-01317a1d9485 9a505de2215d44919cd3f3867df24bfb d0796a19572f46d9a1a2de06fab9db68 - - default default] Security group member updated ['ebbb614f-b60c-4f3e-8d62-ebbe70950da7']#033[00m Nov 27 05:09:04 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:09:04 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:09:04 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:09:04 localhost podman[325392]: 2025-11-27 10:09:04.394667629 +0000 UTC m=+0.072549875 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:09:04 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:09:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:09:04 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:09:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v369: 177 pgs: 177 active+clean; 146 MiB data, 894 MiB used, 41 GiB / 42 GiB avail Nov 27 05:09:05 localhost nova_compute[284026]: 2025-11-27 10:09:05.636 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "format": "json"}]: dispatch Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:9d5352b6-61be-40ff-9aa8-279df080d01f, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:9d5352b6-61be-40ff-9aa8-279df080d01f, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:06 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '9d5352b6-61be-40ff-9aa8-279df080d01f' of type subvolume Nov 27 05:09:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:06.235+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '9d5352b6-61be-40ff-9aa8-279df080d01f' of type subvolume Nov 27 05:09:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "9d5352b6-61be-40ff-9aa8-279df080d01f", "force": true, "format": "json"}]: dispatch Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/9d5352b6-61be-40ff-9aa8-279df080d01f'' moved to trashcan Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:09:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:9d5352b6-61be-40ff-9aa8-279df080d01f, vol_name:cephfs) < "" Nov 27 05:09:06 localhost nova_compute[284026]: 2025-11-27 10:09:06.586 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v370: 177 pgs: 177 active+clean; 146 MiB data, 898 MiB used, 41 GiB / 42 GiB avail; 62 KiB/s rd, 17 KiB/s wr, 89 op/s Nov 27 05:09:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:07 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/139608806' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:07 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/139608806' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #37. Immutable memtables: 0. Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.582142) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 19] Flushing memtable with next log file: 37 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147582225, "job": 19, "event": "flush_started", "num_memtables": 1, "num_entries": 976, "num_deletes": 254, "total_data_size": 2338447, "memory_usage": 2377480, "flush_reason": "Manual Compaction"} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 19] Level-0 flush table #38: started Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147595210, "cf_name": "default", "job": 19, "event": "table_file_creation", "file_number": 38, "file_size": 1436673, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 23068, "largest_seqno": 24038, "table_properties": {"data_size": 1432450, "index_size": 1822, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1349, "raw_key_size": 11353, "raw_average_key_size": 21, "raw_value_size": 1423353, "raw_average_value_size": 2753, "num_data_blocks": 79, "num_entries": 517, "num_filter_entries": 517, "num_deletions": 254, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238110, "oldest_key_time": 1764238110, "file_creation_time": 1764238147, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 38, "seqno_to_time_mapping": "N/A"}} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 19] Flush lasted 13123 microseconds, and 6058 cpu microseconds. Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.595270) [db/flush_job.cc:967] [default] [JOB 19] Level-0 flush table #38: 1436673 bytes OK Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.595298) [db/memtable_list.cc:519] [default] Level-0 commit table #38 started Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.598107) [db/memtable_list.cc:722] [default] Level-0 commit table #38: memtable #1 done Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.598133) EVENT_LOG_v1 {"time_micros": 1764238147598126, "job": 19, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.598155) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 19] Try to delete WAL files size 2333329, prev total WAL file size 2333653, number of live WAL files 2. Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000034.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.599066) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6D6772737461740034303035' seq:72057594037927935, type:22 .. '6D6772737461740034323536' seq:0, type:0; will stop at (end) Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 20] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 19 Base level 0, inputs: [38(1403KB)], [36(17MB)] Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147599142, "job": 20, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [38], "files_L6": [36], "score": -1, "input_data_size": 20238899, "oldest_snapshot_seqno": -1} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 20] Generated table #39: 12664 keys, 18223126 bytes, temperature: kUnknown Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147727123, "cf_name": "default", "job": 20, "event": "table_file_creation", "file_number": 39, "file_size": 18223126, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 18153020, "index_size": 37365, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 31685, "raw_key_size": 340702, "raw_average_key_size": 26, "raw_value_size": 17939356, "raw_average_value_size": 1416, "num_data_blocks": 1399, "num_entries": 12664, "num_filter_entries": 12664, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238147, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 39, "seqno_to_time_mapping": "N/A"}} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.727839) [db/compaction/compaction_job.cc:1663] [default] [JOB 20] Compacted 1@0 + 1@6 files to L6 => 18223126 bytes Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.730125) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 157.9 rd, 142.2 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(1.4, 17.9 +0.0 blob) out(17.4 +0.0 blob), read-write-amplify(26.8) write-amplify(12.7) OK, records in: 13182, records dropped: 518 output_compression: NoCompression Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.730168) EVENT_LOG_v1 {"time_micros": 1764238147730149, "job": 20, "event": "compaction_finished", "compaction_time_micros": 128190, "compaction_time_cpu_micros": 54382, "output_level": 6, "num_output_files": 1, "total_output_size": 18223126, "num_input_records": 13182, "num_output_records": 12664, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000038.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147730644, "job": 20, "event": "table_file_deletion", "file_number": 38} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000036.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238147734355, "job": 20, "event": "table_file_deletion", "file_number": 36} Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.598929) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.734483) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.734491) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.734493) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.734496) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:09:07.734497) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:09:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e175 e175: 6 total, 6 up, 6 in Nov 27 05:09:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e175 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:08 localhost podman[242678]: time="2025-11-27T10:09:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:09:08 localhost podman[242678]: @ - - [27/Nov/2025:10:09:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159340 "" "Go-http-client/1.1" Nov 27 05:09:08 localhost podman[242678]: @ - - [27/Nov/2025:10:09:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20211 "" "Go-http-client/1.1" Nov 27 05:09:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v372: 177 pgs: 177 active+clean; 146 MiB data, 898 MiB used, 41 GiB / 42 GiB avail; 56 KiB/s rd, 15 KiB/s wr, 80 op/s Nov 27 05:09:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:09.576 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '18'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:09:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:09:09 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:09:10 localhost podman[325414]: 2025-11-27 10:09:10.003852611 +0000 UTC m=+0.089896361 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, release=1755695350, maintainer=Red Hat, Inc., container_name=openstack_network_exporter, io.openshift.expose-services=, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, version=9.6, com.redhat.component=ubi9-minimal-container, config_id=edpm, io.openshift.tags=minimal rhel9, vcs-type=git, url=https://catalog.redhat.com/en/search?searchType=containers, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 05:09:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:09:10 localhost podman[325413]: 2025-11-27 10:09:10.059094649 +0000 UTC m=+0.147673278 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:09:10 localhost podman[325413]: 2025-11-27 10:09:10.067917496 +0000 UTC m=+0.156496155 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:09:10 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:09:10 localhost podman[325414]: 2025-11-27 10:09:10.119819954 +0000 UTC m=+0.205863714 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-type=git, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, release=1755695350, name=ubi9-minimal, container_name=openstack_network_exporter) Nov 27 05:09:10 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:09:10 localhost podman[325443]: 2025-11-27 10:09:10.219504918 +0000 UTC m=+0.190698996 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:09:10 localhost podman[325443]: 2025-11-27 10:09:10.238869439 +0000 UTC m=+0.210063507 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute) Nov 27 05:09:10 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:09:10 localhost nova_compute[284026]: 2025-11-27 10:09:10.639 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v373: 177 pgs: 177 active+clean; 146 MiB data, 903 MiB used, 41 GiB / 42 GiB avail; 77 KiB/s rd, 20 KiB/s wr, 108 op/s Nov 27 05:09:11 localhost nova_compute[284026]: 2025-11-27 10:09:11.624 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e176 e176: 6 total, 6 up, 6 in Nov 27 05:09:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e176 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v375: 177 pgs: 177 active+clean; 146 MiB data, 903 MiB used, 41 GiB / 42 GiB avail; 77 KiB/s rd, 20 KiB/s wr, 108 op/s Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v376: 177 pgs: 177 active+clean; 146 MiB data, 903 MiB used, 41 GiB / 42 GiB avail; 30 KiB/s rd, 7.9 KiB/s wr, 41 op/s Nov 27 05:09:15 localhost nova_compute[284026]: 2025-11-27 10:09:15.642 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:16 localhost nova_compute[284026]: 2025-11-27 10:09:16.624 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v377: 177 pgs: 177 active+clean; 146 MiB data, 883 MiB used, 41 GiB / 42 GiB avail; 27 KiB/s rd, 11 KiB/s wr, 39 op/s Nov 27 05:09:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e176 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:09:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:09:18 localhost podman[325477]: 2025-11-27 10:09:18.997488563 +0000 UTC m=+0.085142434 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_id=ovn_controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:09:19 localhost podman[325478]: 2025-11-27 10:09:19.072925164 +0000 UTC m=+0.156237798 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:09:19 localhost podman[325477]: 2025-11-27 10:09:19.085945404 +0000 UTC m=+0.173599285 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:09:19 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:09:19 localhost podman[325478]: 2025-11-27 10:09:19.136479515 +0000 UTC m=+0.219792189 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:09:19 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:09:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v378: 177 pgs: 177 active+clean; 146 MiB data, 883 MiB used, 41 GiB / 42 GiB avail; 25 KiB/s rd, 10 KiB/s wr, 36 op/s Nov 27 05:09:19 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:19.720 265123 INFO neutron.agent.linux.ip_lib [None req-a466abdf-61ec-4375-98f2-dda9408afef3 - - - - - -] Device tap4c510cc6-18 cannot be used as it has no MAC address#033[00m Nov 27 05:09:19 localhost nova_compute[284026]: 2025-11-27 10:09:19.747 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:19 localhost kernel: device tap4c510cc6-18 entered promiscuous mode Nov 27 05:09:19 localhost NetworkManager[5971]: [1764238159.7598] manager: (tap4c510cc6-18): new Generic device (/org/freedesktop/NetworkManager/Devices/47) Nov 27 05:09:19 localhost ovn_controller[156436]: 2025-11-27T10:09:19Z|00269|binding|INFO|Claiming lport 4c510cc6-1843-43bd-837b-d8727d46ce1a for this chassis. Nov 27 05:09:19 localhost ovn_controller[156436]: 2025-11-27T10:09:19Z|00270|binding|INFO|4c510cc6-1843-43bd-837b-d8727d46ce1a: Claiming unknown Nov 27 05:09:19 localhost nova_compute[284026]: 2025-11-27 10:09:19.763 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:19 localhost systemd-udevd[325535]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:09:19 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:19.773 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-65f4b000-5412-4adf-a4ea-160ba8daccb8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-65f4b000-5412-4adf-a4ea-160ba8daccb8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=a0e46088-446a-45e6-9d90-ca2d0db958d0, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4c510cc6-1843-43bd-837b-d8727d46ce1a) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:19 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:19.776 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4c510cc6-1843-43bd-837b-d8727d46ce1a in datapath 65f4b000-5412-4adf-a4ea-160ba8daccb8 bound to our chassis#033[00m Nov 27 05:09:19 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:19.777 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 65f4b000-5412-4adf-a4ea-160ba8daccb8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:19 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:19.778 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[cf0986af-8d8b-4e64-945d-effe488f54d4]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost ovn_controller[156436]: 2025-11-27T10:09:19Z|00271|binding|INFO|Setting lport 4c510cc6-1843-43bd-837b-d8727d46ce1a ovn-installed in OVS Nov 27 05:09:19 localhost ovn_controller[156436]: 2025-11-27T10:09:19Z|00272|binding|INFO|Setting lport 4c510cc6-1843-43bd-837b-d8727d46ce1a up in Southbound Nov 27 05:09:19 localhost nova_compute[284026]: 2025-11-27 10:09:19.807 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost journal[232028]: ethtool ioctl error on tap4c510cc6-18: No such device Nov 27 05:09:19 localhost nova_compute[284026]: 2025-11-27 10:09:19.846 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:19 localhost nova_compute[284026]: 2025-11-27 10:09:19.877 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:20 localhost nova_compute[284026]: 2025-11-27 10:09:20.645 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:20 localhost podman[325606]: Nov 27 05:09:20 localhost podman[325606]: 2025-11-27 10:09:20.882042896 +0000 UTC m=+0.108127802 container create 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:09:20 localhost systemd[1]: Started libpod-conmon-0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291.scope. Nov 27 05:09:20 localhost podman[325606]: 2025-11-27 10:09:20.831464094 +0000 UTC m=+0.057549060 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:09:20 localhost systemd[1]: Started libcrun container. Nov 27 05:09:20 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/bfc7b963fa9f56a8b74ccda8f2f049b8b658cbc4e6aeb598b7cbd13e9bd1dcbd/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:09:20 localhost podman[325606]: 2025-11-27 10:09:20.984172886 +0000 UTC m=+0.210257792 container init 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:09:20 localhost podman[325606]: 2025-11-27 10:09:20.99585226 +0000 UTC m=+0.221937166 container start 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:09:21 localhost dnsmasq[325624]: started, version 2.85 cachesize 150 Nov 27 05:09:21 localhost dnsmasq[325624]: DNS service limited to local subnets Nov 27 05:09:21 localhost dnsmasq[325624]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:09:21 localhost dnsmasq[325624]: warning: no upstream servers configured Nov 27 05:09:21 localhost dnsmasq-dhcp[325624]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:09:21 localhost dnsmasq[325624]: read /var/lib/neutron/dhcp/65f4b000-5412-4adf-a4ea-160ba8daccb8/addn_hosts - 0 addresses Nov 27 05:09:21 localhost dnsmasq-dhcp[325624]: read /var/lib/neutron/dhcp/65f4b000-5412-4adf-a4ea-160ba8daccb8/host Nov 27 05:09:21 localhost dnsmasq-dhcp[325624]: read /var/lib/neutron/dhcp/65f4b000-5412-4adf-a4ea-160ba8daccb8/opts Nov 27 05:09:21 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:21.198 265123 INFO neutron.agent.dhcp.agent [None req-61a362d3-5d6d-4fe0-8599-2c4fd3f023e2 - - - - - -] DHCP configuration for ports {'3da5697e-748b-4f6b-bafe-cab08d5d8dfe'} is completed#033[00m Nov 27 05:09:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v379: 177 pgs: 177 active+clean; 146 MiB data, 883 MiB used, 41 GiB / 42 GiB avail; 26 KiB/s rd, 5.5 KiB/s wr, 37 op/s Nov 27 05:09:21 localhost nova_compute[284026]: 2025-11-27 10:09:21.666 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:21 localhost nova_compute[284026]: 2025-11-27 10:09:21.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e176 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v380: 177 pgs: 177 active+clean; 146 MiB data, 883 MiB used, 41 GiB / 42 GiB avail; 24 KiB/s rd, 5.2 KiB/s wr, 35 op/s Nov 27 05:09:23 localhost nova_compute[284026]: 2025-11-27 10:09:23.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:23 localhost nova_compute[284026]: 2025-11-27 10:09:23.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:24 localhost nova_compute[284026]: 2025-11-27 10:09:24.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:24 localhost nova_compute[284026]: 2025-11-27 10:09:24.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:24 localhost nova_compute[284026]: 2025-11-27 10:09:24.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:09:25 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:25.198 265123 INFO neutron.agent.linux.ip_lib [None req-a6ff36f1-e862-49de-a474-b6f478f3400a - - - - - -] Device tap29bac02a-a9 cannot be used as it has no MAC address#033[00m Nov 27 05:09:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v381: 177 pgs: 177 active+clean; 146 MiB data, 883 MiB used, 41 GiB / 42 GiB avail; 22 KiB/s rd, 4.6 KiB/s wr, 31 op/s Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.267 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost kernel: device tap29bac02a-a9 entered promiscuous mode Nov 27 05:09:25 localhost NetworkManager[5971]: [1764238165.2747] manager: (tap29bac02a-a9): new Generic device (/org/freedesktop/NetworkManager/Devices/48) Nov 27 05:09:25 localhost ovn_controller[156436]: 2025-11-27T10:09:25Z|00273|binding|INFO|Claiming lport 29bac02a-a96d-47b8-80de-f7872ae8104b for this chassis. Nov 27 05:09:25 localhost ovn_controller[156436]: 2025-11-27T10:09:25Z|00274|binding|INFO|29bac02a-a96d-47b8-80de-f7872ae8104b: Claiming unknown Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.276 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost systemd-udevd[325635]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:09:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:25.286 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::3/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-440a04da-a6c3-4b64-81bc-55a7d84134e8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-440a04da-a6c3-4b64-81bc-55a7d84134e8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=d91363fb-5d09-4e0d-b1df-49fda785c120, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=29bac02a-a96d-47b8-80de-f7872ae8104b) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:25.288 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 29bac02a-a96d-47b8-80de-f7872ae8104b in datapath 440a04da-a6c3-4b64-81bc-55a7d84134e8 bound to our chassis#033[00m Nov 27 05:09:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:25.288 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 440a04da-a6c3-4b64-81bc-55a7d84134e8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:25 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:25.290 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[10c88f0e-efc4-4261-892d-5d408d64770e]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost ovn_controller[156436]: 2025-11-27T10:09:25Z|00275|binding|INFO|Setting lport 29bac02a-a96d-47b8-80de-f7872ae8104b ovn-installed in OVS Nov 27 05:09:25 localhost ovn_controller[156436]: 2025-11-27T10:09:25Z|00276|binding|INFO|Setting lport 29bac02a-a96d-47b8-80de-f7872ae8104b up in Southbound Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.312 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost journal[232028]: ethtool ioctl error on tap29bac02a-a9: No such device Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.363 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.395 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost nova_compute[284026]: 2025-11-27 10:09:25.647 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:25 localhost openstack_network_exporter[244641]: ERROR 10:09:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:09:25 localhost openstack_network_exporter[244641]: ERROR 10:09:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:09:25 localhost openstack_network_exporter[244641]: ERROR 10:09:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:09:25 localhost openstack_network_exporter[244641]: ERROR 10:09:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:09:25 localhost openstack_network_exporter[244641]: Nov 27 05:09:25 localhost openstack_network_exporter[244641]: ERROR 10:09:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:09:25 localhost openstack_network_exporter[244641]: Nov 27 05:09:26 localhost ovn_controller[156436]: 2025-11-27T10:09:26Z|00277|binding|INFO|Removing iface tap29bac02a-a9 ovn-installed in OVS Nov 27 05:09:26 localhost ovn_controller[156436]: 2025-11-27T10:09:26Z|00278|binding|INFO|Removing lport 29bac02a-a96d-47b8-80de-f7872ae8104b ovn-installed in OVS Nov 27 05:09:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:26.174 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 78212d5d-e4ca-4dd0-a072-a56bbed7be38 with type ""#033[00m Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.174 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:26.175 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-440a04da-a6c3-4b64-81bc-55a7d84134e8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-440a04da-a6c3-4b64-81bc-55a7d84134e8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=d91363fb-5d09-4e0d-b1df-49fda785c120, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=29bac02a-a96d-47b8-80de-f7872ae8104b) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:26.178 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 29bac02a-a96d-47b8-80de-f7872ae8104b in datapath 440a04da-a6c3-4b64-81bc-55a7d84134e8 unbound from our chassis#033[00m Nov 27 05:09:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:26.179 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 440a04da-a6c3-4b64-81bc-55a7d84134e8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:26 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:26.180 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[651d5410-a7ea-4c45-a2ca-6aad0455f729]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.181 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost podman[325706]: Nov 27 05:09:26 localhost podman[325706]: 2025-11-27 10:09:26.31594346 +0000 UTC m=+0.094855425 container create 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:09:26 localhost systemd[1]: Started libpod-conmon-6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b.scope. Nov 27 05:09:26 localhost ovn_controller[156436]: 2025-11-27T10:09:26Z|00279|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:09:26 localhost podman[325706]: 2025-11-27 10:09:26.274097263 +0000 UTC m=+0.053009238 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:09:26 localhost systemd[1]: Started libcrun container. Nov 27 05:09:26 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6b80699433729054de5b6300ae7b786ecec30d82e4ce69bc1c7eab75f21ac048/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:09:26 localhost podman[325706]: 2025-11-27 10:09:26.396586952 +0000 UTC m=+0.175498907 container init 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:09:26 localhost podman[325706]: 2025-11-27 10:09:26.406401945 +0000 UTC m=+0.185313910 container start 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.409 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost dnsmasq[325724]: started, version 2.85 cachesize 150 Nov 27 05:09:26 localhost dnsmasq[325724]: DNS service limited to local subnets Nov 27 05:09:26 localhost dnsmasq[325724]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:09:26 localhost dnsmasq[325724]: warning: no upstream servers configured Nov 27 05:09:26 localhost dnsmasq-dhcp[325724]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:09:26 localhost dnsmasq[325724]: read /var/lib/neutron/dhcp/440a04da-a6c3-4b64-81bc-55a7d84134e8/addn_hosts - 0 addresses Nov 27 05:09:26 localhost dnsmasq-dhcp[325724]: read /var/lib/neutron/dhcp/440a04da-a6c3-4b64-81bc-55a7d84134e8/host Nov 27 05:09:26 localhost dnsmasq-dhcp[325724]: read /var/lib/neutron/dhcp/440a04da-a6c3-4b64-81bc-55a7d84134e8/opts Nov 27 05:09:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:26.544 265123 INFO neutron.agent.dhcp.agent [None req-cd7c76c4-7cc3-4dba-96a4-1d15f3d75215 - - - - - -] DHCP configuration for ports {'0c163bf9-3b0b-454e-b0dd-bc71095f4447'} is completed#033[00m Nov 27 05:09:26 localhost dnsmasq[325724]: exiting on receipt of SIGTERM Nov 27 05:09:26 localhost podman[325742]: 2025-11-27 10:09:26.654759363 +0000 UTC m=+0.060136210 container kill 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:09:26 localhost systemd[1]: libpod-6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b.scope: Deactivated successfully. Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.670 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost podman[325754]: 2025-11-27 10:09:26.717942025 +0000 UTC m=+0.051219531 container died 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:09:26 localhost podman[325754]: 2025-11-27 10:09:26.750495841 +0000 UTC m=+0.083773337 container cleanup 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:09:26 localhost systemd[1]: libpod-conmon-6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b.scope: Deactivated successfully. Nov 27 05:09:26 localhost podman[325756]: 2025-11-27 10:09:26.813127158 +0000 UTC m=+0.132908351 container remove 6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-440a04da-a6c3-4b64-81bc-55a7d84134e8, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2) Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.827 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost kernel: device tap29bac02a-a9 left promiscuous mode Nov 27 05:09:26 localhost nova_compute[284026]: 2025-11-27 10:09:26.840 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:26.875 265123 INFO neutron.agent.dhcp.agent [None req-674c416e-d763-4951-b149-ce7f3bb3a0f0 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:26.875 265123 INFO neutron.agent.dhcp.agent [None req-674c416e-d763-4951-b149-ce7f3bb3a0f0 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v382: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 42 KiB/s rd, 5.7 KiB/s wr, 58 op/s Nov 27 05:09:27 localhost systemd[1]: var-lib-containers-storage-overlay-6b80699433729054de5b6300ae7b786ecec30d82e4ce69bc1c7eab75f21ac048-merged.mount: Deactivated successfully. Nov 27 05:09:27 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-6af47a85752baae3d0fa17dbcf3df1ff6649c34bb84f5c50793af217501ce78b-userdata-shm.mount: Deactivated successfully. Nov 27 05:09:27 localhost systemd[1]: run-netns-qdhcp\x2d440a04da\x2da6c3\x2d4b64\x2d81bc\x2d55a7d84134e8.mount: Deactivated successfully. Nov 27 05:09:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e176 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:28 localhost nova_compute[284026]: 2025-11-27 10:09:28.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:28 localhost nova_compute[284026]: 2025-11-27 10:09:28.729 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v383: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 41 KiB/s rd, 2.2 KiB/s wr, 55 op/s Nov 27 05:09:29 localhost dnsmasq[324862]: exiting on receipt of SIGTERM Nov 27 05:09:29 localhost podman[325800]: 2025-11-27 10:09:29.572941538 +0000 UTC m=+0.062479524 container kill 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:09:29 localhost systemd[1]: libpod-39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e.scope: Deactivated successfully. Nov 27 05:09:29 localhost podman[325813]: 2025-11-27 10:09:29.653040345 +0000 UTC m=+0.067926440 container died 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:09:29 localhost systemd[1]: tmp-crun.3xuhqH.mount: Deactivated successfully. Nov 27 05:09:29 localhost ovn_controller[156436]: 2025-11-27T10:09:29Z|00280|binding|INFO|Removing iface tapc98b1720-02 ovn-installed in OVS Nov 27 05:09:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:29.672 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 5ecf16fb-f118-4ea6-8e39-f349ca8744a1 with type ""#033[00m Nov 27 05:09:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:29.673 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.255.242/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-bd45c14b-d688-4b39-bb9e-8d0539609d09', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-bd45c14b-d688-4b39-bb9e-8d0539609d09', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '04d6be087105405690248faa72ceeb41', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=df5df163-ea1e-4879-b1cb-bdc0a447121e, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=c98b1720-0211-488c-964d-21d0073c26d1) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:29.674 162092 INFO neutron.agent.ovn.metadata.agent [-] Port c98b1720-0211-488c-964d-21d0073c26d1 in datapath bd45c14b-d688-4b39-bb9e-8d0539609d09 unbound from our chassis#033[00m Nov 27 05:09:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:29.676 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network bd45c14b-d688-4b39-bb9e-8d0539609d09, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:09:29 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:29.677 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b434df96-fb8f-46d1-b7f6-01727decb1d0]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:29 localhost ovn_controller[156436]: 2025-11-27T10:09:29Z|00281|binding|INFO|Removing lport c98b1720-0211-488c-964d-21d0073c26d1 ovn-installed in OVS Nov 27 05:09:29 localhost nova_compute[284026]: 2025-11-27 10:09:29.684 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:29 localhost podman[325813]: 2025-11-27 10:09:29.69226031 +0000 UTC m=+0.107146355 container cleanup 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:09:29 localhost systemd[1]: libpod-conmon-39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e.scope: Deactivated successfully. Nov 27 05:09:29 localhost podman[325818]: 2025-11-27 10:09:29.734641802 +0000 UTC m=+0.137685819 container remove 39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-bd45c14b-d688-4b39-bb9e-8d0539609d09, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:09:29 localhost nova_compute[284026]: 2025-11-27 10:09:29.746 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:29 localhost kernel: device tapc98b1720-02 left promiscuous mode Nov 27 05:09:29 localhost nova_compute[284026]: 2025-11-27 10:09:29.766 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:29 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:29.814 265123 INFO neutron.agent.dhcp.agent [None req-d0bfdfb1-63f7-4df4-83e5-9bfb1334dc03 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:30 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:30.033 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:30 localhost ovn_controller[156436]: 2025-11-27T10:09:30Z|00282|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:09:30 localhost nova_compute[284026]: 2025-11-27 10:09:30.413 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:30 localhost systemd[1]: var-lib-containers-storage-overlay-02d8eedafbedc7999cbf7efe0c4caaa867535d7079c0a9ed9f3d7cb5fdc2db1b-merged.mount: Deactivated successfully. Nov 27 05:09:30 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-39c034428051b8f0ba59a0694f16da4eec9726e15fc7e8ab91bf35376186470e-userdata-shm.mount: Deactivated successfully. Nov 27 05:09:30 localhost systemd[1]: run-netns-qdhcp\x2dbd45c14b\x2dd688\x2d4b39\x2dbb9e\x2d8d0539609d09.mount: Deactivated successfully. Nov 27 05:09:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e177 e177: 6 total, 6 up, 6 in Nov 27 05:09:30 localhost nova_compute[284026]: 2025-11-27 10:09:30.649 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v385: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 27 KiB/s rd, 1.7 KiB/s wr, 37 op/s Nov 27 05:09:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e178 e178: 6 total, 6 up, 6 in Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.711 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:31.892 265123 INFO neutron.agent.linux.ip_lib [None req-d1ed04d2-8f1f-4baa-969a-67266c154e98 - - - - - -] Device tap2aaf3198-28 cannot be used as it has no MAC address#033[00m Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.918 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost kernel: device tap2aaf3198-28 entered promiscuous mode Nov 27 05:09:31 localhost NetworkManager[5971]: [1764238171.9263] manager: (tap2aaf3198-28): new Generic device (/org/freedesktop/NetworkManager/Devices/49) Nov 27 05:09:31 localhost ovn_controller[156436]: 2025-11-27T10:09:31Z|00283|binding|INFO|Claiming lport 2aaf3198-28b0-4016-8303-d626c6bac5bd for this chassis. Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.926 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost ovn_controller[156436]: 2025-11-27T10:09:31Z|00284|binding|INFO|2aaf3198-28b0-4016-8303-d626c6bac5bd: Claiming unknown Nov 27 05:09:31 localhost systemd-udevd[325853]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:09:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:31.945 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=215d0d9a-8e0f-4b1b-9b26-40d206f7faba, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=2aaf3198-28b0-4016-8303-d626c6bac5bd) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:31.947 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 2aaf3198-28b0-4016-8303-d626c6bac5bd in datapath d137fc7e-bdfc-4c44-b2dd-d0095319f0e6 bound to our chassis#033[00m Nov 27 05:09:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:31.948 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network d137fc7e-bdfc-4c44-b2dd-d0095319f0e6 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:31.949 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3643e549-a3d6-48d3-abda-901d700fe19a]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost ovn_controller[156436]: 2025-11-27T10:09:31Z|00285|binding|INFO|Setting lport 2aaf3198-28b0-4016-8303-d626c6bac5bd ovn-installed in OVS Nov 27 05:09:31 localhost ovn_controller[156436]: 2025-11-27T10:09:31Z|00286|binding|INFO|Setting lport 2aaf3198-28b0-4016-8303-d626c6bac5bd up in Southbound Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.960 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.963 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost journal[232028]: ethtool ioctl error on tap2aaf3198-28: No such device Nov 27 05:09:31 localhost nova_compute[284026]: 2025-11-27 10:09:31.995 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.024 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta' Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "format": "json"}]: dispatch Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:32 localhost ovn_controller[156436]: 2025-11-27T10:09:32Z|00287|binding|INFO|Removing iface tap2aaf3198-28 ovn-installed in OVS Nov 27 05:09:32 localhost ovn_controller[156436]: 2025-11-27T10:09:32Z|00288|binding|INFO|Removing lport 2aaf3198-28b0-4016-8303-d626c6bac5bd ovn-installed in OVS Nov 27 05:09:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:32.657 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 725d54c4-e3da-4962-ab7b-0cd132220a23 with type ""#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.659 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:32.661 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=215d0d9a-8e0f-4b1b-9b26-40d206f7faba, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=2aaf3198-28b0-4016-8303-d626c6bac5bd) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:32.664 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 2aaf3198-28b0-4016-8303-d626c6bac5bd in datapath d137fc7e-bdfc-4c44-b2dd-d0095319f0e6 unbound from our chassis#033[00m Nov 27 05:09:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:32.664 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network d137fc7e-bdfc-4c44-b2dd-d0095319f0e6 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:32 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:32.665 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[19f07c7b-bc54-4eaf-aa68-f24d26bb3704]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.666 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.749 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.749 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.750 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.750 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.750 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:09:32 localhost ovn_controller[156436]: 2025-11-27T10:09:32Z|00289|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:09:32 localhost podman[325922]: Nov 27 05:09:32 localhost podman[325922]: 2025-11-27 10:09:32.841729944 +0000 UTC m=+0.093445518 container create 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:09:32 localhost podman[325922]: 2025-11-27 10:09:32.798413967 +0000 UTC m=+0.050129551 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:09:32 localhost nova_compute[284026]: 2025-11-27 10:09:32.922 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:09:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:09:32 localhost systemd[1]: Started libpod-conmon-30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1.scope. Nov 27 05:09:32 localhost systemd[1]: Started libcrun container. Nov 27 05:09:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/dcbb414372855ac4a1434b094757d7626f74ad15660b804bcad26eca485ec22d/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:09:32 localhost podman[325922]: 2025-11-27 10:09:32.986641025 +0000 UTC m=+0.238356599 container init 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 05:09:33 localhost podman[325922]: 2025-11-27 10:09:33.005854542 +0000 UTC m=+0.257570116 container start 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:09:33 localhost dnsmasq[325986]: started, version 2.85 cachesize 150 Nov 27 05:09:33 localhost dnsmasq[325986]: DNS service limited to local subnets Nov 27 05:09:33 localhost dnsmasq[325986]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:09:33 localhost dnsmasq[325986]: warning: no upstream servers configured Nov 27 05:09:33 localhost dnsmasq-dhcp[325986]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:09:33 localhost dnsmasq[325986]: read /var/lib/neutron/dhcp/d137fc7e-bdfc-4c44-b2dd-d0095319f0e6/addn_hosts - 0 addresses Nov 27 05:09:33 localhost dnsmasq-dhcp[325986]: read /var/lib/neutron/dhcp/d137fc7e-bdfc-4c44-b2dd-d0095319f0e6/host Nov 27 05:09:33 localhost dnsmasq-dhcp[325986]: read /var/lib/neutron/dhcp/d137fc7e-bdfc-4c44-b2dd-d0095319f0e6/opts Nov 27 05:09:33 localhost podman[325947]: 2025-11-27 10:09:33.003094988 +0000 UTC m=+0.067673703 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:09:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e178 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:33 localhost podman[325948]: 2025-11-27 10:09:33.070096082 +0000 UTC m=+0.129694663 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:09:33 localhost podman[325948]: 2025-11-27 10:09:33.080860542 +0000 UTC m=+0.140459083 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 05:09:33 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:09:33 localhost podman[325947]: 2025-11-27 10:09:33.136262534 +0000 UTC m=+0.200841279 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:09:33 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:09:33 localhost dnsmasq[325986]: exiting on receipt of SIGTERM Nov 27 05:09:33 localhost podman[326014]: 2025-11-27 10:09:33.225813986 +0000 UTC m=+0.059030601 container kill 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:09:33 localhost systemd[1]: libpod-30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1.scope: Deactivated successfully. Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.249 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.499s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:09:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v387: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 34 KiB/s rd, 2.1 KiB/s wr, 46 op/s Nov 27 05:09:33 localhost podman[326029]: 2025-11-27 10:09:33.300955648 +0000 UTC m=+0.058295500 container died 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:09:33 localhost podman[326029]: 2025-11-27 10:09:33.328778907 +0000 UTC m=+0.086118719 container cleanup 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:09:33 localhost systemd[1]: libpod-conmon-30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1.scope: Deactivated successfully. Nov 27 05:09:33 localhost podman[326032]: 2025-11-27 10:09:33.383866221 +0000 UTC m=+0.135842138 container remove 30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-d137fc7e-bdfc-4c44-b2dd-d0095319f0e6, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.397 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:33 localhost kernel: device tap2aaf3198-28 left promiscuous mode Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.412 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.451 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.451 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:09:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:33.630 265123 INFO neutron.agent.dhcp.agent [None req-d7e7ae05-bc3d-45f5-ad0e-fcd9c59122ef - - - - - -] DHCP configuration for ports {'86b0c6e2-8af4-4ecd-b7a7-41c06cb0e724'} is completed#033[00m Nov 27 05:09:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:33.632 265123 INFO neutron.agent.dhcp.agent [None req-8b378ddf-6de8-45a5-a556-0fdbef61d544 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:33.633 265123 INFO neutron.agent.dhcp.agent [None req-8b378ddf-6de8-45a5-a556-0fdbef61d544 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.675 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.677 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11133MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.677 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.677 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:09:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e179 e179: 6 total, 6 up, 6 in Nov 27 05:09:33 localhost systemd[1]: var-lib-containers-storage-overlay-dcbb414372855ac4a1434b094757d7626f74ad15660b804bcad26eca485ec22d-merged.mount: Deactivated successfully. Nov 27 05:09:33 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-30e0d4c2896b0c9ea2e2223ea8bfbe84839f11f75870a5e7b441a12f2ecfbeb1-userdata-shm.mount: Deactivated successfully. Nov 27 05:09:33 localhost systemd[1]: run-netns-qdhcp\x2dd137fc7e\x2dbdfc\x2d4c44\x2db2dd\x2dd0095319f0e6.mount: Deactivated successfully. Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.887 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.888 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.888 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:09:33 localhost nova_compute[284026]: 2025-11-27 10:09:33.975 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:09:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:09:34 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/164972887' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:09:34 localhost nova_compute[284026]: 2025-11-27 10:09:34.440 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.466s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:09:34 localhost nova_compute[284026]: 2025-11-27 10:09:34.447 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:09:34 localhost nova_compute[284026]: 2025-11-27 10:09:34.464 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:09:34 localhost nova_compute[284026]: 2025-11-27 10:09:34.467 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:09:34 localhost nova_compute[284026]: 2025-11-27 10:09:34.467 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.790s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:09:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e180 e180: 6 total, 6 up, 6 in Nov 27 05:09:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v390: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail Nov 27 05:09:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "snap_name": "4faf3386-a11b-466c-a025-408a3f4f9e7f", "format": "json"}]: dispatch Nov 27 05:09:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:09:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:35 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/598517682' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:35 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/598517682' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:35 localhost nova_compute[284026]: 2025-11-27 10:09:35.652 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.468 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.468 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.469 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.560 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.560 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.561 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.561 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:09:36 localhost nova_compute[284026]: 2025-11-27 10:09:36.712 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:37 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2142350764' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:37 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2142350764' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:37 localhost nova_compute[284026]: 2025-11-27 10:09:37.157 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:09:37 localhost nova_compute[284026]: 2025-11-27 10:09:37.179 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:09:37 localhost nova_compute[284026]: 2025-11-27 10:09:37.179 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:09:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v391: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 67 KiB/s rd, 12 KiB/s wr, 92 op/s Nov 27 05:09:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:37 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3038329243' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:37 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3038329243' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e181 e181: 6 total, 6 up, 6 in Nov 27 05:09:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e181 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:38 localhost podman[242678]: time="2025-11-27T10:09:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:09:38 localhost podman[242678]: @ - - [27/Nov/2025:10:09:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159329 "" "Go-http-client/1.1" Nov 27 05:09:38 localhost podman[242678]: @ - - [27/Nov/2025:10:09:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20216 "" "Go-http-client/1.1" Nov 27 05:09:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot clone", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "snap_name": "4faf3386-a11b-466c-a025-408a3f4f9e7f", "target_sub_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "format": "json"}]: dispatch Nov 27 05:09:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, target_sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:09:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 273 bytes to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' Nov 27 05:09:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] tracking-id 6adf9ef2-cc37-42c0-b4a3-526766dec89d for path b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 246 bytes to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, target_sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:09:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "format": "json"}]: dispatch Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.035+0000 7f54e6161640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.035+0000 7f54e6161640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.035+0000 7f54e6161640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.035+0000 7f54e6161640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.035+0000 7f54e6161640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] cloning to subvolume path: /volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58 Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] starting clone: (cephfs, None, a087a1df-d6f4-418c-ac4c-417b86d57a58) Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.065+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.065+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.065+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.065+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:39.065+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] Delayed cloning (cephfs, None, a087a1df-d6f4-418c-ac4c-417b86d57a58) -- by 0 seconds Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 277 bytes to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' Nov 27 05:09:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta' Nov 27 05:09:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v393: 177 pgs: 177 active+clean; 146 MiB data, 867 MiB used, 41 GiB / 42 GiB avail; 67 KiB/s rd, 12 KiB/s wr, 92 op/s Nov 27 05:09:39 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:39.883 2 INFO neutron.agent.securitygroups_rpc [None req-5b124142-053a-45ac-938c-9354503213ee cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['bb3531e8-d08c-4788-b8f7-7fdc03f17536']#033[00m Nov 27 05:09:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:40.051 2 INFO neutron.agent.securitygroups_rpc [None req-c43b393c-5cd8-4e58-a7bb-ddd5dbcdf946 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['bb3531e8-d08c-4788-b8f7-7fdc03f17536']#033[00m Nov 27 05:09:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:40 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4099354662' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:40 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4099354662' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:40 localhost nova_compute[284026]: 2025-11-27 10:09:40.654 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e182 e182: 6 total, 6 up, 6 in Nov 27 05:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:09:40 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:09:41 localhost podman[326108]: 2025-11-27 10:09:41.014315899 +0000 UTC m=+0.098183465 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, maintainer=Red Hat, Inc., architecture=x86_64, io.openshift.tags=minimal rhel9, name=ubi9-minimal, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, container_name=openstack_network_exporter, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 05:09:41 localhost podman[326106]: 2025-11-27 10:09:40.990339783 +0000 UTC m=+0.082421770 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible) Nov 27 05:09:41 localhost podman[326108]: 2025-11-27 10:09:41.055821457 +0000 UTC m=+0.139689003 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, vcs-type=git, container_name=openstack_network_exporter, distribution-scope=public, version=9.6, build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, maintainer=Red Hat, Inc., managed_by=edpm_ansible, config_id=edpm, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.buildah.version=1.33.7, architecture=x86_64, name=ubi9-minimal, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container) Nov 27 05:09:41 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:09:41 localhost podman[326106]: 2025-11-27 10:09:41.072851675 +0000 UTC m=+0.164933692 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=edpm, io.buildah.version=1.41.3, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:09:41 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:09:41 localhost podman[326107]: 2025-11-27 10:09:41.058252052 +0000 UTC m=+0.147103802 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:09:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:09:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:41 localhost podman[326107]: 2025-11-27 10:09:41.137962538 +0000 UTC m=+0.226814308 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:09:41 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:09:41 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:41.228 2 INFO neutron.agent.securitygroups_rpc [None req-bc0f7c49-ce82-47a3-9efd-718126503228 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v395: 177 pgs: 177 active+clean; 146 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 172 KiB/s rd, 33 KiB/s wr, 237 op/s Nov 27 05:09:41 localhost nova_compute[284026]: 2025-11-27 10:09:41.753 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:41 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:41.787 2 INFO neutron.agent.securitygroups_rpc [None req-72ea8ab3-301b-4fa2-b440-22518ffe5cff cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:42.118 2 INFO neutron.agent.securitygroups_rpc [None req-f9ab5a09-4772-478a-900d-faaa7111dbb0 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] copying data from b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.snap/4faf3386-a11b-466c-a025-408a3f4f9e7f/ae8a0512-151e-4985-b302-1105fbfc18f6' to b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/1dc04d26-a8a4-4631-94d1-2e36afd4e215' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "format": "json"}]: dispatch Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 274 bytes to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] untracking 6adf9ef2-cc37-42c0-b4a3-526766dec89d Nov 27 05:09:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:42.383 2 INFO neutron.agent.securitygroups_rpc [None req-401abc30-9c1e-43b0-86d3-f126a13f4470 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 151 bytes to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta.tmp' to config b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58/.meta' Nov 27 05:09:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] finished clone: (cephfs, None, a087a1df-d6f4-418c-ac4c-417b86d57a58) Nov 27 05:09:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:42.529 2 INFO neutron.agent.securitygroups_rpc [None req-4181f5ec-799b-4cb9-bcbb-241d98784c2e cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e183 e183: 6 total, 6 up, 6 in Nov 27 05:09:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:42.738 2 INFO neutron.agent.securitygroups_rpc [None req-f5b40bc1-4885-49e5-a83f-9ea0574ee169 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:42.978 2 INFO neutron.agent.securitygroups_rpc [None req-3a181878-e7bf-4d35-b181-e614b0e259ba cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e183 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:43 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:43.141 2 INFO neutron.agent.securitygroups_rpc [None req-08bcdbd4-8604-45da-a7b0-fa189cd1e29c cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v397: 177 pgs: 177 active+clean; 146 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 121 KiB/s rd, 24 KiB/s wr, 166 op/s Nov 27 05:09:43 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:43.293 2 INFO neutron.agent.securitygroups_rpc [None req-7f37d32c-9595-41af-b8a6-80c039d10e12 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:43.577 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:09:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:43.578 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:09:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:43.578 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:09:43 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:43.807 2 INFO neutron.agent.securitygroups_rpc [None req-187165ba-8006-4943-9c89-d8d4cf170059 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['685dfb95-0c12-4ece-95ec-cd34b266ca3b']#033[00m Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.138 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.139 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.143 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9ee09cc0-d58e-472c-b198-2546f0db2e2d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.140071', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3227d7ba-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '172a404b8a8dd51ddb3ba8ddf5c946f725bf4393df7af7b3c17f9f249616c9db'}]}, 'timestamp': '2025-11-27 10:09:44.144188', '_unique_id': 'f909e5e924e34e8cbf74853d274332d0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.145 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.147 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.159 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.159 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd1f2133f-5137-4713-9230-8cb980fda36a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.147376', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '322a3e42-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': 'e8ac4d9e427113456bd7c47b9dc09920c01cec34dcabf722c152201b320f6410'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.147376', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '322a51b6-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': '1ddbf2c2b6cb4ebf1fca9e872f11850cf27fba5f5ead28084194e18efdb7b60e'}]}, 'timestamp': '2025-11-27 10:09:44.160298', '_unique_id': '691805b515a4427b90828a07284c5d81'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.161 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.162 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.162 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.163 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b547aac5-d797-4312-a0b5-28b705e1ef76', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.162763', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '322ac326-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': '3ea1b46dcd03acf63c32f43c3d8e3ed230fadc51c7cc65a252b20750202dfe15'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.162763', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '322ad2f8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': '7a959ae50743be84217e2da5b9091c2ea726a20a465f78b9405042765c2b846b'}]}, 'timestamp': '2025-11-27 10:09:44.163701', '_unique_id': '79a7e8d428044e2f8ef70f9b888f477f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.164 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.165 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'db2edb74-0db0-4c2e-a939-052944a1a684', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.166021', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '322b4274-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '47b6644b1507155c1a6d48a3d9205705bb4fbd3b70939b74f437c0d43dfe9872'}]}, 'timestamp': '2025-11-27 10:09:44.166492', '_unique_id': 'd26b0d48f3234bff9c8bf01055428443'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.167 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.169 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.186 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 17380000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'da18d183-3d70-4bea-9b82-f2071088daad', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 17380000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:09:44.169756', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '322e7070-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.37466679, 'message_signature': '99e60c21f9f31b772201b90c21ecb2fc6430dfc90aea7f76a1b6c86b7b8a1d1d'}]}, 'timestamp': '2025-11-27 10:09:44.187322', '_unique_id': '80eade15f9a846798d75e8edac6277e2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.189 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.218 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b03acbfb-4e83-41b0-ab8c-bb75785f43f9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.190101', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '32335fae-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': 'da8dc112d88faa33a4657828a09a01135f0e604952aee98cd8cd225d450236fe'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.190101', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '32338d30-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '19edcfecc00174fbac7e144953c11ff00e0ad8d3e7713afc2f255be1a7caafcb'}]}, 'timestamp': '2025-11-27 10:09:44.220870', '_unique_id': 'c12763e842cf4f6aa1435aa9177d179e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.225 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '022c8cdb-b9c8-49d9-84f4-a1eccab5af42', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.226313', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '32348b7c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '170b7b8b862ad6d1ad0d746b7157103fda8bfa825d7ac7806070260d25f92b72'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.226313', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3234c47a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': 'e0a0f6d99f47127ad80f086e2ab9a3c44184be26fa6de784bf17d116970c391c'}]}, 'timestamp': '2025-11-27 10:09:44.228825', '_unique_id': 'eda125eb775b49009bfe3c2cd81c708e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.232 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.232 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '48749a62-3608-4e23-b750-5735d3c79aec', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.232793', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3235769a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '16ec5124c12df4fd3cd6fdf63ef466e946cfa24f06be025bc6fffb1633820a87'}]}, 'timestamp': '2025-11-27 10:09:44.233369', '_unique_id': '67de5fa3aae143f8b0ec8a5bddf60692'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.237 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6d5d8d5e-6c57-4a69-bd77-65691ff80283', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:09:44.237029', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '32362040-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.37466679, 'message_signature': 'a8a7ba4ac11507be9a1434e94e0e3b02d666eb15f2a9d91ca835434fbf6154d1'}]}, 'timestamp': '2025-11-27 10:09:44.237777', '_unique_id': '469fcaeaba0c49b0a9a1cb571a510868'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.240 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.240 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2a7adf17-9491-406d-923b-b5527587d82d', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.241143', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3236bb68-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': 'e922fe469d1c8421aaffef53cdff17daab51dfd1311cdcc998c0c49573574225'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.241143', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3236d026-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.335437674, 'message_signature': 'c05e0ddddb9b13dc06d6cfa0a48fba260da2d63916dad0fc444b39f58f01c576'}]}, 'timestamp': '2025-11-27 10:09:44.242181', '_unique_id': 'b4e9cd78c3ec4064a535697df97c315b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.244 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '570e5afd-7c0b-45ac-936b-5b246dd0e5de', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.245004', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3237532a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '66124ce9bf066371924e9ac608fded59cf8802f0a4f08845cd0e817ec95bf203'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.245004', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '32376a18-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '6b3280c6a77b705df3e8c2545115fb23cc55de23e0e7af97be64a502076b6368'}]}, 'timestamp': '2025-11-27 10:09:44.246129', '_unique_id': '3a56cab520294bf8be18905187e2a11d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.247 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.248 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.249 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'fe2ef209-7162-479b-b683-c974571eebc0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.248849', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '3237e8a8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '8994550df37ff00597740f1c376af700bf56b40ee159954ef28d68eed198ebe0'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.248849', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '3237fc8a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '64127deb50aa489d324ee58c6a7cc1fe5084b99b36cd776a086b870609b94c9f'}]}, 'timestamp': '2025-11-27 10:09:44.249875', '_unique_id': '456f82dc53044c5c8b0cd91b4d36fccb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.250 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.252 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.252 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3378ca62-1c68-414b-bcfd-c8dcc7dceef9', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.252658', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '32387db8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '9c072ca0b3f8ab03eab3908bedf598a9f26d70910562d5a6aa0fbb215b80ae86'}]}, 'timestamp': '2025-11-27 10:09:44.253207', '_unique_id': 'fe047432ce844afd959bea242e044137'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.255 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f7860745-ca60-49f4-b3b1-70f5d7621cf6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.256082', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '323902e2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '804ffe6dd17e2ca0040a412e8f7925e97b71d2adc3e14a973b1b839ab94a3cad'}]}, 'timestamp': '2025-11-27 10:09:44.256646', '_unique_id': '3d2caf54dc2e459684ecdf00ecf5209c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '87092461-01bc-43b4-8ff4-7fe21eac167c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.259382', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '323980e6-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '51ac37aaa01a71cbdbd1eac094f519ad29094f896b4d0ee9d964f61ec0cf78e0'}]}, 'timestamp': '2025-11-27 10:09:44.259754', '_unique_id': '099c4a53055a4ffca0fca33f7e9cdeff'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.261 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '47a4af95-016e-446a-bc9a-0aeccd8625ca', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.261947', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '3239e3d8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': 'f93a1848b26bb65931013069804216eed70d89cb77d04974a3b5e40da8f87c08'}]}, 'timestamp': '2025-11-27 10:09:44.262288', '_unique_id': '3a37e249b1524c1abe0586b48d28aa23'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.263 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '86890b2f-eed0-4330-82dc-f5c552987cfe', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.264092', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '323a3798-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': 'afb3e4dd0df8185d474406a6a01313bff694c88bd7040a96b2a00259de46404d'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.264092', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '323a44a4-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '297bf159910567461994fd08057ca811046a5224cc0166643045d5ca3904a298'}]}, 'timestamp': '2025-11-27 10:09:44.264749', '_unique_id': '8b4c304a9ef840c78f89c9081bf1dde9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.266 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.266 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.266 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'aa448046-d281-476d-8f48-0d1e7e5f27be', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.266586', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '323a9a6c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '5e1f66dfac97866cde0ea7dc64cd2153cb0e3675dc5db62aa0999692906bb02d'}]}, 'timestamp': '2025-11-27 10:09:44.266961', '_unique_id': '12fc55a0c5044af88c903933abf30f8e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.268 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.268 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.269 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6f2f9481-b7d4-47b2-b9a5-d3e3158acc5c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:09:44.268823', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '323af02a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': '5d619e54c89b7e0f4d2b968376a4fb7f5328e3d67996752d3e99bf12bbdcef4f'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:09:44.268823', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '323afbc4-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.378161575, 'message_signature': 'cb9a0c08d4c6c409b01b693656c232188904f0322b02d206a8e6ad3aac20c9ae'}]}, 'timestamp': '2025-11-27 10:09:44.269436', '_unique_id': '63a0e28fe04a47c6bdfa5985eb5c177f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.270 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7a4b01c2-2dba-43d6-9282-b4d53601636e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.271100', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '323b4926-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': '1f8ddd5832e5156e6e8c9c135d99d277729b9f68bde3c5b3b197fb7a49990c83'}]}, 'timestamp': '2025-11-27 10:09:44.271433', '_unique_id': 'ef8b91802bd549e68250c921cf161b03'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.272 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '02092dc1-4d26-4430-ab82-290685ed6bf0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:09:44.273040', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '323b94f8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12414.328163118, 'message_signature': 'c4b6a3b9e54aa1eb518bf85897831bb0ffb95442efb0ef4f91c067582fa6d52c'}]}, 'timestamp': '2025-11-27 10:09:44.273374', '_unique_id': '4923bd33f2f949e4a12806a0e21d007d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.273 12 ERROR oslo_messaging.notify.messaging Nov 27 05:09:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:09:44.274 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:09:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:09:44 Nov 27 05:09:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:09:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:09:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['vms', 'backups', 'manila_data', 'images', 'manila_metadata', 'volumes', '.mgr'] Nov 27 05:09:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:09:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:44.489 2 INFO neutron.agent.securitygroups_rpc [None req-d76aa726-a37b-4184-b34a-ddb6e4e79a31 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['682855b0-743e-49fc-928d-c786a9263da1']#033[00m Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:09:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "snap_name": "adb912c7-d74c-4153-bfac-0a7bdfecd982", "format": "json"}]: dispatch Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 3.271566164154104e-06 of space, bias 1.0, pg target 0.0006532227107761028 quantized to 32 (current 32) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299928461753211 of space, bias 1.0, pg target 0.8585523828633911 quantized to 32 (current 32) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.0905220547180346e-06 of space, bias 1.0, pg target 0.00021701388888888888 quantized to 32 (current 32) Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:09:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 3.107987855946399e-05 of space, bias 4.0, pg target 0.024739583333333332 quantized to 16 (current 16) Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:09:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:09:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v398: 177 pgs: 177 active+clean; 146 MiB data, 868 MiB used, 41 GiB / 42 GiB avail; 97 KiB/s rd, 19 KiB/s wr, 133 op/s Nov 27 05:09:45 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:45.629 2 INFO neutron.agent.securitygroups_rpc [None req-b0fe9ae9-2341-4552-ae48-fd70284253ae cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['2d8a84f8-b042-4e67-9eb9-239da8db6ac8']#033[00m Nov 27 05:09:45 localhost nova_compute[284026]: 2025-11-27 10:09:45.657 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:45 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:45.772 2 INFO neutron.agent.securitygroups_rpc [None req-a93d5db3-8303-46ef-9f66-f21e21dcdacb cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['2d8a84f8-b042-4e67-9eb9-239da8db6ac8']#033[00m Nov 27 05:09:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e184 e184: 6 total, 6 up, 6 in Nov 27 05:09:46 localhost ceph-mgr[290377]: [devicehealth INFO root] Check health Nov 27 05:09:46 localhost nova_compute[284026]: 2025-11-27 10:09:46.783 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:46 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1547256362' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:46 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1547256362' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:47 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:47.160 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=19, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=18) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:47 localhost nova_compute[284026]: 2025-11-27 10:09:47.161 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:47 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:47.162 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 5 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:09:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v400: 177 pgs: 177 active+clean; 146 MiB data, 869 MiB used, 41 GiB / 42 GiB avail; 18 KiB/s rd, 24 KiB/s wr, 28 op/s Nov 27 05:09:47 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:47.755 2 INFO neutron.agent.securitygroups_rpc [None req-7da6cc8a-2059-43e5-88ac-56d629f72ee6 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['593903e2-6abb-46c7-b0f4-a4a488366487']#033[00m Nov 27 05:09:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e184 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:48 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:48.237 2 INFO neutron.agent.securitygroups_rpc [None req-a5a0cbcf-f2b0-48d0-ab70-80add6c32d79 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['593903e2-6abb-46c7-b0f4-a4a488366487']#033[00m Nov 27 05:09:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:09:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4104832685' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:09:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:09:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/4104832685' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:09:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "snap_name": "adb912c7-d74c-4153-bfac-0a7bdfecd982_24837742-b609-48c6-af99-056350db6973", "force": true, "format": "json"}]: dispatch Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982_24837742-b609-48c6-af99-056350db6973, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta' Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982_24837742-b609-48c6-af99-056350db6973, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v401: 177 pgs: 177 active+clean; 146 MiB data, 869 MiB used, 41 GiB / 42 GiB avail; 14 KiB/s rd, 19 KiB/s wr, 23 op/s Nov 27 05:09:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "snap_name": "adb912c7-d74c-4153-bfac-0a7bdfecd982", "force": true, "format": "json"}]: dispatch Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta.tmp' to config b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857/.meta' Nov 27 05:09:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:adb912c7-d74c-4153-bfac-0a7bdfecd982, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:49 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:49.686 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:09:49Z, description=, device_id=ddab7444-4618-4651-9096-3c66a0c9252f, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e01bf965-2abd-4d93-9dec-ba64871eaa44, ip_allocation=immediate, mac_address=fa:16:3e:83:99:16, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=2501, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:09:49Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:09:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e185 e185: 6 total, 6 up, 6 in Nov 27 05:09:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:49.696 2 INFO neutron.agent.securitygroups_rpc [None req-f61675a8-3801-4cf3-a4ee-120f91662220 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:09:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:09:49 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:09:49 localhost podman[326182]: 2025-11-27 10:09:49.974693707 +0000 UTC m=+0.055883296 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:09:49 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:09:49 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:09:50 localhost podman[326179]: 2025-11-27 10:09:50.050012275 +0000 UTC m=+0.138871270 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_controller) Nov 27 05:09:50 localhost podman[326180]: 2025-11-27 10:09:50.107673988 +0000 UTC m=+0.192999458 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:09:50 localhost podman[326180]: 2025-11-27 10:09:50.120697659 +0000 UTC m=+0.206023099 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:09:50 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:09:50 localhost podman[326179]: 2025-11-27 10:09:50.175827023 +0000 UTC m=+0.264686058 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller) Nov 27 05:09:50 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:09:50 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:50.215 2 INFO neutron.agent.securitygroups_rpc [None req-8c68a1d4-98fd-4933-b025-811616c84715 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:50 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:50.307 265123 INFO neutron.agent.dhcp.agent [None req-287bbfbc-be81-4a5a-bf9f-795081132c80 - - - - - -] DHCP configuration for ports {'e01bf965-2abd-4d93-9dec-ba64871eaa44'} is completed#033[00m Nov 27 05:09:50 localhost nova_compute[284026]: 2025-11-27 10:09:50.659 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v403: 177 pgs: 177 active+clean; 258 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 14 MiB/s wr, 116 op/s Nov 27 05:09:51 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:51.585 2 INFO neutron.agent.securitygroups_rpc [None req-5e0d5755-1a0b-40de-a347-958bbb4bd698 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e186 e186: 6 total, 6 up, 6 in Nov 27 05:09:51 localhost nova_compute[284026]: 2025-11-27 10:09:51.822 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:51 localhost nova_compute[284026]: 2025-11-27 10:09:51.851 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:52 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:52.010 2 INFO neutron.agent.securitygroups_rpc [None req-dd439b12-ac80-4be0-b9fb-7239b8ac476a cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:52 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:52.164 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '19'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:09:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "format": "json"}]: dispatch Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:c20b9d42-4601-4748-a46a-eb38c58f9857, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:c20b9d42-4601-4748-a46a-eb38c58f9857, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:09:52 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'c20b9d42-4601-4748-a46a-eb38c58f9857' of type subvolume Nov 27 05:09:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:09:52.571+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'c20b9d42-4601-4748-a46a-eb38c58f9857' of type subvolume Nov 27 05:09:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "c20b9d42-4601-4748-a46a-eb38c58f9857", "force": true, "format": "json"}]: dispatch Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/c20b9d42-4601-4748-a46a-eb38c58f9857'' moved to trashcan Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:09:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:c20b9d42-4601-4748-a46a-eb38c58f9857, vol_name:cephfs) < "" Nov 27 05:09:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e187 e187: 6 total, 6 up, 6 in Nov 27 05:09:52 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:52.661 2 INFO neutron.agent.securitygroups_rpc [None req-fe29d790-9910-4b1c-afc6-78269a68ef32 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e187 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v406: 177 pgs: 177 active+clean; 258 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 87 KiB/s rd, 19 MiB/s wr, 124 op/s Nov 27 05:09:53 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:53.344 2 INFO neutron.agent.securitygroups_rpc [None req-6ec0f8d3-1b16-46e1-90cd-fd0033c67426 cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['0e6f25f9-b3eb-4a32-ba5f-c9ff506ae0a6']#033[00m Nov 27 05:09:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e188 e188: 6 total, 6 up, 6 in Nov 27 05:09:54 localhost neutron_sriov_agent[258162]: 2025-11-27 10:09:54.602 2 INFO neutron.agent.securitygroups_rpc [None req-ee551b57-0d61-4eae-97a4-bc099233f7ae cdb52d3c355340e99c53e7a6716fcec2 0c9016260e42419b8e1360d8e2445571 - - default default] Security group rule updated ['3131b977-c89b-49f6-a3fe-479882bfaf9d']#033[00m Nov 27 05:09:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v408: 177 pgs: 177 active+clean; 258 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 94 KiB/s rd, 20 MiB/s wr, 134 op/s Nov 27 05:09:55 localhost nova_compute[284026]: 2025-11-27 10:09:55.663 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:55 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:09:55 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2959903710' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:09:55 localhost openstack_network_exporter[244641]: ERROR 10:09:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:09:55 localhost openstack_network_exporter[244641]: ERROR 10:09:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:09:55 localhost openstack_network_exporter[244641]: ERROR 10:09:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:09:55 localhost openstack_network_exporter[244641]: ERROR 10:09:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:09:55 localhost openstack_network_exporter[244641]: Nov 27 05:09:55 localhost openstack_network_exporter[244641]: ERROR 10:09:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:09:55 localhost openstack_network_exporter[244641]: Nov 27 05:09:56 localhost nova_compute[284026]: 2025-11-27 10:09:56.861 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e189 e189: 6 total, 6 up, 6 in Nov 27 05:09:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v410: 177 pgs: 177 active+clean; 146 MiB data, 895 MiB used, 41 GiB / 42 GiB avail; 90 KiB/s rd, 21 KiB/s wr, 129 op/s Nov 27 05:09:57 localhost nova_compute[284026]: 2025-11-27 10:09:57.432 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e190 e190: 6 total, 6 up, 6 in Nov 27 05:09:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e190 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:09:58 localhost dnsmasq[325624]: exiting on receipt of SIGTERM Nov 27 05:09:58 localhost podman[326260]: 2025-11-27 10:09:58.336794064 +0000 UTC m=+0.053312477 container kill 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:09:58 localhost systemd[1]: libpod-0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291.scope: Deactivated successfully. Nov 27 05:09:58 localhost ovn_controller[156436]: 2025-11-27T10:09:58Z|00290|binding|INFO|Removing iface tap4c510cc6-18 ovn-installed in OVS Nov 27 05:09:58 localhost ovn_controller[156436]: 2025-11-27T10:09:58Z|00291|binding|INFO|Removing lport 4c510cc6-1843-43bd-837b-d8727d46ce1a ovn-installed in OVS Nov 27 05:09:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:58.420 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port 721e0d67-3e5d-42aa-85e9-a424e8e90570 with type ""#033[00m Nov 27 05:09:58 localhost podman[326274]: 2025-11-27 10:09:58.421656989 +0000 UTC m=+0.064497287 container died 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:09:58 localhost nova_compute[284026]: 2025-11-27 10:09:58.421 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:58.422 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-65f4b000-5412-4adf-a4ea-160ba8daccb8', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-65f4b000-5412-4adf-a4ea-160ba8daccb8', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '1f9c676ffabb494696e987a548f13bbd', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=a0e46088-446a-45e6-9d90-ca2d0db958d0, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4c510cc6-1843-43bd-837b-d8727d46ce1a) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:09:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:58.424 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4c510cc6-1843-43bd-837b-d8727d46ce1a in datapath 65f4b000-5412-4adf-a4ea-160ba8daccb8 unbound from our chassis#033[00m Nov 27 05:09:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:58.426 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 65f4b000-5412-4adf-a4ea-160ba8daccb8 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:09:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:09:58.427 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[2f101cd0-00c0-4b7a-acf0-6f60ea941936]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:09:58 localhost nova_compute[284026]: 2025-11-27 10:09:58.430 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:58 localhost systemd[1]: tmp-crun.bHWKyz.mount: Deactivated successfully. Nov 27 05:09:58 localhost podman[326274]: 2025-11-27 10:09:58.521412925 +0000 UTC m=+0.164253193 container cleanup 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:09:58 localhost systemd[1]: libpod-conmon-0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291.scope: Deactivated successfully. Nov 27 05:09:58 localhost podman[326275]: 2025-11-27 10:09:58.544354652 +0000 UTC m=+0.181230750 container remove 0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-65f4b000-5412-4adf-a4ea-160ba8daccb8, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:09:58 localhost nova_compute[284026]: 2025-11-27 10:09:58.554 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:58 localhost kernel: device tap4c510cc6-18 left promiscuous mode Nov 27 05:09:58 localhost nova_compute[284026]: 2025-11-27 10:09:58.574 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:58.590 265123 INFO neutron.agent.dhcp.agent [None req-4f9b933e-6571-4e1c-bd70-a54295191f3b - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:09:58.594 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:09:58 localhost ovn_controller[156436]: 2025-11-27T10:09:58Z|00292|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:09:58 localhost nova_compute[284026]: 2025-11-27 10:09:58.839 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:09:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v412: 177 pgs: 177 active+clean; 146 MiB data, 895 MiB used, 41 GiB / 42 GiB avail; 83 KiB/s rd, 20 KiB/s wr, 119 op/s Nov 27 05:09:59 localhost systemd[1]: var-lib-containers-storage-overlay-bfc7b963fa9f56a8b74ccda8f2f049b8b658cbc4e6aeb598b7cbd13e9bd1dcbd-merged.mount: Deactivated successfully. Nov 27 05:09:59 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-0cf2f016f9edadd6e4d39d2e4503017638f2f3b8be1423ce894420f402760291-userdata-shm.mount: Deactivated successfully. Nov 27 05:09:59 localhost systemd[1]: run-netns-qdhcp\x2d65f4b000\x2d5412\x2d4adf\x2da4ea\x2d160ba8daccb8.mount: Deactivated successfully. Nov 27 05:09:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e191 e191: 6 total, 6 up, 6 in Nov 27 05:10:00 localhost nova_compute[284026]: 2025-11-27 10:10:00.666 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:00 localhost ceph-mon[305520]: overall HEALTH_WARN 1 stray daemon(s) not managed by cephadm; 1 stray host(s) with 1 daemon(s) not managed by cephadm Nov 27 05:10:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v414: 177 pgs: 177 active+clean; 146 MiB data, 913 MiB used, 41 GiB / 42 GiB avail; 3.5 MiB/s rd, 28 KiB/s wr, 203 op/s Nov 27 05:10:01 localhost nova_compute[284026]: 2025-11-27 10:10:01.900 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e192 e192: 6 total, 6 up, 6 in Nov 27 05:10:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e193 e193: 6 total, 6 up, 6 in Nov 27 05:10:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e193 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v417: 177 pgs: 177 active+clean; 146 MiB data, 913 MiB used, 41 GiB / 42 GiB avail; 3.9 MiB/s rd, 9.7 KiB/s wr, 93 op/s Nov 27 05:10:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:10:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:10:03 localhost systemd[1]: tmp-crun.Z5TE7D.mount: Deactivated successfully. Nov 27 05:10:03 localhost systemd[1]: tmp-crun.nWS1HZ.mount: Deactivated successfully. Nov 27 05:10:03 localhost podman[326320]: 2025-11-27 10:10:03.907101509 +0000 UTC m=+0.124426372 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, managed_by=edpm_ansible, container_name=multipathd, io.buildah.version=1.41.3) Nov 27 05:10:03 localhost podman[326320]: 2025-11-27 10:10:03.919866762 +0000 UTC m=+0.137191625 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, container_name=multipathd, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:10:03 localhost podman[326318]: 2025-11-27 10:10:03.878024275 +0000 UTC m=+0.097706751 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 05:10:03 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:10:03 localhost podman[326318]: 2025-11-27 10:10:03.96397887 +0000 UTC m=+0.183661326 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:10:03 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:10:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e194 e194: 6 total, 6 up, 6 in Nov 27 05:10:04 localhost podman[326442]: 2025-11-27 10:10:04.821174221 +0000 UTC m=+0.097043774 container exec 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, description=Red Hat Ceph Storage 7, maintainer=Guillaume Abrioux , build-date=2025-09-24T08:57:55, GIT_BRANCH=main, vcs-type=git, GIT_CLEAN=True, com.redhat.component=rhceph-container, name=rhceph, io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.buildah.version=1.33.12, architecture=x86_64, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, release=553, RELEASE=main, io.openshift.tags=rhceph ceph, version=7, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, com.redhat.license_terms=https://www.redhat.com/agreements, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9) Nov 27 05:10:04 localhost podman[326442]: 2025-11-27 10:10:04.910978319 +0000 UTC m=+0.186847922 container exec_died 062e3e32d1c76dce719af8e518fb20e1bcbb923b641a3c5c2c673abec3822f73 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-crash-np0005537446, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, GIT_CLEAN=True, architecture=x86_64, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., RELEASE=main, distribution-scope=public, description=Red Hat Ceph Storage 7, io.openshift.expose-services=, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , GIT_BRANCH=main, release=553, CEPH_POINT_RELEASE=, ceph=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., io.k8s.description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, version=7, build-date=2025-09-24T08:57:55, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e195 e195: 6 total, 6 up, 6 in Nov 27 05:10:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v420: 177 pgs: 177 active+clean; 146 MiB data, 913 MiB used, 41 GiB / 42 GiB avail Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 05:10:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 05:10:05 localhost nova_compute[284026]: 2025-11-27 10:10:05.669 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "mon dump", "format": "json"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2002411310' entity='client.openstack' cmd={"prefix": "mon dump", "format": "json"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm INFO root] Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [INF] : Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config set, name=osd_memory_target}] v 0) Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mgr[290377]: [cephadm WARNING cephadm.serve] Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mgr[290377]: log_channel(cephadm) log [WRN] : Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:10:06 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 2476560a-63e4-48f0-a84e-56924473d0c5 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:10:06 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 2476560a-63e4-48f0-a84e-56924473d0c5 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:10:06 localhost ceph-mgr[290377]: [progress INFO root] Completed event 2476560a-63e4-48f0-a84e-56924473d0c5 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:10:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:10:06 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:10:06 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:06.548 2 INFO neutron.agent.securitygroups_rpc [None req-05fa11db-7976-4810-be86-e9cf01768aa0 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:06 localhost nova_compute[284026]: 2025-11-27 10:10:06.939 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537445.localdomain to 836.6M Nov 27 05:10:07 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537445.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537446.localdomain to 836.6M Nov 27 05:10:07 localhost ceph-mon[305520]: Adjusting osd_memory_target on np0005537444.localdomain to 836.6M Nov 27 05:10:07 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537446.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:07 localhost ceph-mon[305520]: Unable to set osd_memory_target on np0005537444.localdomain to 877246668: error parsing value: Value '877246668' is below minimum 939524096 Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch Nov 27 05:10:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v421: 177 pgs: 177 active+clean; 192 MiB data, 955 MiB used, 41 GiB / 42 GiB avail; 143 KiB/s rd, 4.0 MiB/s wr, 205 op/s Nov 27 05:10:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e196 e196: 6 total, 6 up, 6 in Nov 27 05:10:07 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:07.537 265123 INFO neutron.agent.linux.ip_lib [None req-98f1f5ac-4aa4-4313-905b-7568bd074f40 - - - - - -] Device tap7223af45-8b cannot be used as it has no MAC address#033[00m Nov 27 05:10:07 localhost nova_compute[284026]: 2025-11-27 10:10:07.568 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:07 localhost kernel: device tap7223af45-8b entered promiscuous mode Nov 27 05:10:07 localhost nova_compute[284026]: 2025-11-27 10:10:07.581 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:07 localhost ovn_controller[156436]: 2025-11-27T10:10:07Z|00293|binding|INFO|Claiming lport 7223af45-8bda-4c4b-93f0-864b601af7e7 for this chassis. Nov 27 05:10:07 localhost ovn_controller[156436]: 2025-11-27T10:10:07Z|00294|binding|INFO|7223af45-8bda-4c4b-93f0-864b601af7e7: Claiming unknown Nov 27 05:10:07 localhost NetworkManager[5971]: [1764238207.5862] manager: (tap7223af45-8b): new Generic device (/org/freedesktop/NetworkManager/Devices/50) Nov 27 05:10:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:07.600 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:0:ffff::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-12671eaa-864f-4aab-98f5-21123177cd60', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-12671eaa-864f-4aab-98f5-21123177cd60', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '292341d0c2ae492d9f1c680cc1b88f1e', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c207ff98-6fa5-4c37-ba07-efb5cc85a81c, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=7223af45-8bda-4c4b-93f0-864b601af7e7) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:07.602 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 7223af45-8bda-4c4b-93f0-864b601af7e7 in datapath 12671eaa-864f-4aab-98f5-21123177cd60 bound to our chassis#033[00m Nov 27 05:10:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:07.603 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 12671eaa-864f-4aab-98f5-21123177cd60 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:10:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:07.604 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ac51fb7d-d371-4012-8ca1-22b646522b34]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:07 localhost ovn_controller[156436]: 2025-11-27T10:10:07Z|00295|binding|INFO|Setting lport 7223af45-8bda-4c4b-93f0-864b601af7e7 ovn-installed in OVS Nov 27 05:10:07 localhost ovn_controller[156436]: 2025-11-27T10:10:07Z|00296|binding|INFO|Setting lport 7223af45-8bda-4c4b-93f0-864b601af7e7 up in Southbound Nov 27 05:10:07 localhost nova_compute[284026]: 2025-11-27 10:10:07.627 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:07 localhost nova_compute[284026]: 2025-11-27 10:10:07.664 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:07 localhost nova_compute[284026]: 2025-11-27 10:10:07.694 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e196 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:08 localhost podman[326711]: Nov 27 05:10:08 localhost podman[326711]: 2025-11-27 10:10:08.615254461 +0000 UTC m=+0.091649818 container create 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:10:08 localhost systemd[1]: Started libpod-conmon-72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd.scope. Nov 27 05:10:08 localhost podman[326711]: 2025-11-27 10:10:08.570424514 +0000 UTC m=+0.046819881 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:10:08 localhost systemd[1]: tmp-crun.6YDg89.mount: Deactivated successfully. Nov 27 05:10:08 localhost systemd[1]: Started libcrun container. Nov 27 05:10:08 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/6f81a373d2019bdd63d37cb55863549270978e47b342b29b79eecc8cd4f5fc1e/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:10:08 localhost podman[326711]: 2025-11-27 10:10:08.702027857 +0000 UTC m=+0.178423224 container init 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:10:08 localhost podman[326711]: 2025-11-27 10:10:08.712634583 +0000 UTC m=+0.189029950 container start 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true) Nov 27 05:10:08 localhost dnsmasq[326729]: started, version 2.85 cachesize 150 Nov 27 05:10:08 localhost dnsmasq[326729]: DNS service limited to local subnets Nov 27 05:10:08 localhost dnsmasq[326729]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:10:08 localhost dnsmasq[326729]: warning: no upstream servers configured Nov 27 05:10:08 localhost dnsmasq-dhcp[326729]: DHCPv6, static leases only on 2001:db8:0:ffff::, lease time 1d Nov 27 05:10:08 localhost dnsmasq[326729]: read /var/lib/neutron/dhcp/12671eaa-864f-4aab-98f5-21123177cd60/addn_hosts - 0 addresses Nov 27 05:10:08 localhost dnsmasq-dhcp[326729]: read /var/lib/neutron/dhcp/12671eaa-864f-4aab-98f5-21123177cd60/host Nov 27 05:10:08 localhost dnsmasq-dhcp[326729]: read /var/lib/neutron/dhcp/12671eaa-864f-4aab-98f5-21123177cd60/opts Nov 27 05:10:08 localhost podman[242678]: time="2025-11-27T10:10:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:10:08 localhost podman[242678]: @ - - [27/Nov/2025:10:10:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159336 "" "Go-http-client/1.1" Nov 27 05:10:08 localhost podman[242678]: @ - - [27/Nov/2025:10:10:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20212 "" "Go-http-client/1.1" Nov 27 05:10:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:08.882 265123 INFO neutron.agent.dhcp.agent [None req-77cc0513-be03-4f11-920e-fe3c15f82711 - - - - - -] DHCP configuration for ports {'ec2e5490-82c6-4629-8240-fdcdd807eeb8'} is completed#033[00m Nov 27 05:10:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e197 e197: 6 total, 6 up, 6 in Nov 27 05:10:09 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:09.269 2 INFO neutron.agent.securitygroups_rpc [None req-e6af1a9f-92cb-46b7-99c8-6855b52d029f 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v424: 177 pgs: 177 active+clean; 192 MiB data, 955 MiB used, 41 GiB / 42 GiB avail; 144 KiB/s rd, 4.0 MiB/s wr, 206 op/s Nov 27 05:10:09 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:09.491 2 INFO neutron.agent.securitygroups_rpc [None req-e6af1a9f-92cb-46b7-99c8-6855b52d029f 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:09 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:09.674 2 INFO neutron.agent.securitygroups_rpc [None req-644f7529-ab66-4196-b6ef-ba8fe0204e43 b97eaf918c144b0881c2db1c261ba46b 292341d0c2ae492d9f1c680cc1b88f1e - - default default] Security group member updated ['43be6d15-f9c1-4f5a-a949-2af26325859d']#033[00m Nov 27 05:10:09 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:10:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:10:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:10:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:10.126 2 INFO neutron.agent.securitygroups_rpc [None req-0aed9b3d-cb50-4c49-a630-9164a70a2f28 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e198 e198: 6 total, 6 up, 6 in Nov 27 05:10:10 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:10.162 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:10 localhost nova_compute[284026]: 2025-11-27 10:10:10.673 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:10.767 2 INFO neutron.agent.securitygroups_rpc [None req-d6e02869-aa21-439a-b5ef-c54c477011ad 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v426: 177 pgs: 177 active+clean; 239 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 7.1 MiB/s rd, 7.1 MiB/s wr, 301 op/s Nov 27 05:10:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:10:11 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:11.898 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:10:11 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:10:11 localhost nova_compute[284026]: 2025-11-27 10:10:11.943 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:12 localhost systemd[1]: tmp-crun.mUmEhu.mount: Deactivated successfully. Nov 27 05:10:12 localhost podman[326731]: 2025-11-27 10:10:12.010714848 +0000 UTC m=+0.100043106 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:10:12 localhost podman[326731]: 2025-11-27 10:10:12.025078094 +0000 UTC m=+0.114406362 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:10:12 localhost systemd[1]: tmp-crun.uZpNq2.mount: Deactivated successfully. Nov 27 05:10:12 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:10:12 localhost podman[326732]: 2025-11-27 10:10:12.170559451 +0000 UTC m=+0.251296827 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.component=ubi9-minimal-container, managed_by=edpm_ansible, io.openshift.expose-services=, version=9.6, release=1755695350, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, distribution-scope=public, maintainer=Red Hat, Inc., architecture=x86_64, container_name=openstack_network_exporter, vendor=Red Hat, Inc., vcs-type=git, build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, io.openshift.tags=minimal rhel9) Nov 27 05:10:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e199 e199: 6 total, 6 up, 6 in Nov 27 05:10:12 localhost podman[326730]: 2025-11-27 10:10:12.082236533 +0000 UTC m=+0.172394683 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3) Nov 27 05:10:12 localhost podman[326732]: 2025-11-27 10:10:12.212783758 +0000 UTC m=+0.293521164 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.buildah.version=1.33.7, name=ubi9-minimal, config_id=edpm, release=1755695350, architecture=x86_64, version=9.6, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, maintainer=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, vcs-type=git, distribution-scope=public, build-date=2025-08-20T13:12:41, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:10:12 localhost podman[326730]: 2025-11-27 10:10:12.220987309 +0000 UTC m=+0.311145469 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:10:12 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:10:12 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:10:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e200 e200: 6 total, 6 up, 6 in Nov 27 05:10:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e200 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v429: 177 pgs: 177 active+clean; 239 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 10 MiB/s rd, 5.1 MiB/s wr, 172 op/s Nov 27 05:10:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e201 e201: 6 total, 6 up, 6 in Nov 27 05:10:13 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:13.932 2 INFO neutron.agent.securitygroups_rpc [None req-bc52bb29-f101-48ab-af60-84c4d5951fa9 b97eaf918c144b0881c2db1c261ba46b 292341d0c2ae492d9f1c680cc1b88f1e - - default default] Security group member updated ['43be6d15-f9c1-4f5a-a949-2af26325859d']#033[00m Nov 27 05:10:14 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "format": "json"}]: dispatch Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:10:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v431: 177 pgs: 177 active+clean; 239 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 8.2 MiB/s rd, 4.1 MiB/s wr, 139 op/s Nov 27 05:10:15 localhost nova_compute[284026]: 2025-11-27 10:10:15.675 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e202 e202: 6 total, 6 up, 6 in Nov 27 05:10:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:16.734 2 INFO neutron.agent.securitygroups_rpc [None req-8d5fdc73-5114-44eb-bfcb-512970194fc2 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:16 localhost nova_compute[284026]: 2025-11-27 10:10:16.988 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "format": "json"}]: dispatch Nov 27 05:10:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:10:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:10:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v433: 177 pgs: 177 active+clean; 331 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 4.3 MiB/s rd, 8.3 MiB/s wr, 257 op/s Nov 27 05:10:17 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:17.668 2 INFO neutron.agent.securitygroups_rpc [None req-a5b8e25a-8181-4c00-9396-92dc051d8188 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e202 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:18.360 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:10:17Z, description=, device_id=de9bc218-14ca-496e-8d86-541765a10181, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=caccf5e5-c457-424f-9b18-a9621356f073, ip_allocation=immediate, mac_address=fa:16:3e:87:32:61, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=2648, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:10:18Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:10:18 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:10:18 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:10:18 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:10:18 localhost podman[326805]: 2025-11-27 10:10:18.60556953 +0000 UTC m=+0.061508147 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:10:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e203 e203: 6 total, 6 up, 6 in Nov 27 05:10:18 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:18.876 265123 INFO neutron.agent.dhcp.agent [None req-e39566d2-6cd9-4753-953f-dc61c04ddcfe - - - - - -] DHCP configuration for ports {'caccf5e5-c457-424f-9b18-a9621356f073'} is completed#033[00m Nov 27 05:10:19 localhost nova_compute[284026]: 2025-11-27 10:10:19.245 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v435: 177 pgs: 177 active+clean; 331 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 3.7 MiB/s rd, 7.1 MiB/s wr, 219 op/s Nov 27 05:10:20 localhost nova_compute[284026]: 2025-11-27 10:10:20.678 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:20 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "format": "json"}]: dispatch Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:20 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "a087a1df-d6f4-418c-ac4c-417b86d57a58", "force": true, "format": "json"}]: dispatch Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/a087a1df-d6f4-418c-ac4c-417b86d57a58'' moved to trashcan Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:10:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a087a1df-d6f4-418c-ac4c-417b86d57a58, vol_name:cephfs) < "" Nov 27 05:10:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e204 e204: 6 total, 6 up, 6 in Nov 27 05:10:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:10:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:10:20 localhost podman[326826]: 2025-11-27 10:10:20.995322547 +0000 UTC m=+0.088619078 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 05:10:21 localhost podman[326827]: 2025-11-27 10:10:21.069892254 +0000 UTC m=+0.157722107 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:10:21 localhost podman[326827]: 2025-11-27 10:10:21.080405077 +0000 UTC m=+0.168234960 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:10:21 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:10:21 localhost podman[326826]: 2025-11-27 10:10:21.139213221 +0000 UTC m=+0.232509752 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, container_name=ovn_controller, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:10:21 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:10:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v437: 177 pgs: 177 active+clean; 192 MiB data, 960 MiB used, 41 GiB / 42 GiB avail; 3.8 MiB/s rd, 7.1 MiB/s wr, 371 op/s Nov 27 05:10:22 localhost nova_compute[284026]: 2025-11-27 10:10:22.015 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e205 e205: 6 total, 6 up, 6 in Nov 27 05:10:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e206 e206: 6 total, 6 up, 6 in Nov 27 05:10:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e206 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:23 localhost nova_compute[284026]: 2025-11-27 10:10:23.202 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v440: 177 pgs: 177 active+clean; 192 MiB data, 960 MiB used, 41 GiB / 42 GiB avail; 140 KiB/s rd, 14 KiB/s wr, 202 op/s Nov 27 05:10:23 localhost nova_compute[284026]: 2025-11-27 10:10:23.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:24 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:24.158 2 INFO neutron.agent.securitygroups_rpc [None req-7ed3b03b-e8d5-4325-807a-8f1fa8955167 c86b8f3bf6f9490288bf21849ce8e6b5 67fb135d336b4706a482e380561998d6 - - default default] Security group rule updated ['8afab950-f78e-4af9-961b-2530a2994b49']#033[00m Nov 27 05:10:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:24.199 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=20, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=19) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:24.200 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 10 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:10:24 localhost nova_compute[284026]: 2025-11-27 10:10:24.496 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:24 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:10:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "snap_name": "4faf3386-a11b-466c-a025-408a3f4f9e7f_6d429f8e-002b-4b06-96ea-c77ae67a15aa", "force": true, "format": "json"}]: dispatch Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f_6d429f8e-002b-4b06-96ea-c77ae67a15aa, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:24 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.107:0/586087946' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta' Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f_6d429f8e-002b-4b06-96ea-c77ae67a15aa, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "snap_name": "4faf3386-a11b-466c-a025-408a3f4f9e7f", "force": true, "format": "json"}]: dispatch Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta.tmp' to config b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c/.meta' Nov 27 05:10:24 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:24.601 2 INFO neutron.agent.securitygroups_rpc [None req-70ea95fd-265c-4626-a03a-7ca11498df86 c86b8f3bf6f9490288bf21849ce8e6b5 67fb135d336b4706a482e380561998d6 - - default default] Security group rule updated ['8afab950-f78e-4af9-961b-2530a2994b49']#033[00m Nov 27 05:10:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:4faf3386-a11b-466c-a025-408a3f4f9e7f, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:10:25 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2419308791' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:10:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:10:25 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2419308791' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:10:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v441: 177 pgs: 177 active+clean; 192 MiB data, 960 MiB used, 41 GiB / 42 GiB avail; 105 KiB/s rd, 11 KiB/s wr, 151 op/s Nov 27 05:10:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e207 e207: 6 total, 6 up, 6 in Nov 27 05:10:25 localhost nova_compute[284026]: 2025-11-27 10:10:25.695 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:25 localhost nova_compute[284026]: 2025-11-27 10:10:25.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:25 localhost nova_compute[284026]: 2025-11-27 10:10:25.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:25 localhost nova_compute[284026]: 2025-11-27 10:10:25.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:25 localhost nova_compute[284026]: 2025-11-27 10:10:25.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:10:25 localhost openstack_network_exporter[244641]: ERROR 10:10:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:10:25 localhost openstack_network_exporter[244641]: ERROR 10:10:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:10:25 localhost openstack_network_exporter[244641]: ERROR 10:10:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:10:25 localhost openstack_network_exporter[244641]: ERROR 10:10:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:10:25 localhost openstack_network_exporter[244641]: Nov 27 05:10:25 localhost openstack_network_exporter[244641]: ERROR 10:10:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:10:25 localhost openstack_network_exporter[244641]: Nov 27 05:10:26 localhost nova_compute[284026]: 2025-11-27 10:10:26.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:27 localhost nova_compute[284026]: 2025-11-27 10:10:27.055 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "format": "json"}]: dispatch Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:344f6870-02d2-4889-bf3a-4b255fcad50c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:344f6870-02d2-4889-bf3a-4b255fcad50c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:27 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '344f6870-02d2-4889-bf3a-4b255fcad50c' of type subvolume Nov 27 05:10:27 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:10:27.204+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '344f6870-02d2-4889-bf3a-4b255fcad50c' of type subvolume Nov 27 05:10:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "344f6870-02d2-4889-bf3a-4b255fcad50c", "force": true, "format": "json"}]: dispatch Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/344f6870-02d2-4889-bf3a-4b255fcad50c'' moved to trashcan Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:10:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:344f6870-02d2-4889-bf3a-4b255fcad50c, vol_name:cephfs) < "" Nov 27 05:10:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v443: 177 pgs: 177 active+clean; 146 MiB data, 899 MiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 25 KiB/s wr, 112 op/s Nov 27 05:10:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:10:27 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/336302737' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:10:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:10:27 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/336302737' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:10:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:27.515 2 INFO neutron.agent.securitygroups_rpc [None req-ccc66752-5029-4d52-a4fa-ee329549f8b7 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e208 e208: 6 total, 6 up, 6 in Nov 27 05:10:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e209 e209: 6 total, 6 up, 6 in Nov 27 05:10:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:27.965 2 INFO neutron.agent.securitygroups_rpc [None req-646ff067-7f09-4cb6-b8b2-dfa2d8ad8b01 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e209 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:28 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:28.534 2 INFO neutron.agent.securitygroups_rpc [None req-e4719393-6178-4d06-a744-c290d4b82ceb 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.686 265123 INFO neutron.agent.dhcp.agent [None req-7fcb19df-1928-4d32-ac80-4f52e06fac47 - - - - - -] Synchronizing state#033[00m Nov 27 05:10:28 localhost nova_compute[284026]: 2025-11-27 10:10:28.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.949 265123 INFO neutron.agent.dhcp.agent [None req-6c889203-87cd-436c-8aa9-be9e9385dff3 - - - - - -] All active networks have been fetched through RPC.#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.950 265123 INFO neutron.agent.dhcp.agent [-] Starting network 6859c895-f1d7-49ec-b26e-3f935c4e01e6 dhcp configuration#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.951 265123 INFO neutron.agent.dhcp.agent [-] Finished network 6859c895-f1d7-49ec-b26e-3f935c4e01e6 dhcp configuration#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.951 265123 INFO neutron.agent.dhcp.agent [None req-6c889203-87cd-436c-8aa9-be9e9385dff3 - - - - - -] Synchronizing state complete#033[00m Nov 27 05:10:28 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:28.952 265123 INFO neutron.agent.dhcp.agent [None req-4725bce6-f62d-4867-a1a9-4cff9bbef7d8 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:29 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:29.101 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:29.103 2 INFO neutron.agent.securitygroups_rpc [None req-9d8cbc3a-7efd-4653-9358-f548c4817abf 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v446: 177 pgs: 177 active+clean; 146 MiB data, 899 MiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 25 KiB/s wr, 112 op/s Nov 27 05:10:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:29.495 2 INFO neutron.agent.securitygroups_rpc [None req-34a20abc-d2d5-47f6-8417-c0af16319433 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:29.987 2 INFO neutron.agent.securitygroups_rpc [None req-cb441053-1c88-455d-8fcf-7125e591d025 b97eaf918c144b0881c2db1c261ba46b 292341d0c2ae492d9f1c680cc1b88f1e - - default default] Security group member updated ['43be6d15-f9c1-4f5a-a949-2af26325859d']#033[00m Nov 27 05:10:30 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:30.262 2 INFO neutron.agent.securitygroups_rpc [None req-995ea75a-0b11-4f3a-b55f-147af56239fe 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:30 localhost nova_compute[284026]: 2025-11-27 10:10:30.734 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:31.088 2 INFO neutron.agent.securitygroups_rpc [None req-5272990f-f230-41a5-b21d-629f775f3dc9 b97eaf918c144b0881c2db1c261ba46b 292341d0c2ae492d9f1c680cc1b88f1e - - default default] Security group member updated ['43be6d15-f9c1-4f5a-a949-2af26325859d']#033[00m Nov 27 05:10:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v447: 177 pgs: 177 active+clean; 146 MiB data, 900 MiB used, 41 GiB / 42 GiB avail; 102 KiB/s rd, 44 KiB/s wr, 145 op/s Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.061 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 e210: 6 total, 6 up, 6 in Nov 27 05:10:32 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:32.667 2 INFO neutron.agent.securitygroups_rpc [req-ee6347ba-7efb-4a9c-b736-ddae6eeb2a93 req-f688e23e-a824-4e86-978a-5b9976189e46 c86b8f3bf6f9490288bf21849ce8e6b5 67fb135d336b4706a482e380561998d6 - - default default] Security group member updated ['8afab950-f78e-4af9-961b-2530a2994b49']#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.766 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.766 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.766 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.767 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:10:32 localhost nova_compute[284026]: 2025-11-27 10:10:32.767 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:10:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:10:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2941877206' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.242 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.475s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:10:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v449: 177 pgs: 177 active+clean; 146 MiB data, 900 MiB used, 41 GiB / 42 GiB avail; 23 KiB/s rd, 19 KiB/s wr, 33 op/s Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.311 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.312 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:10:33 localhost dnsmasq[326729]: exiting on receipt of SIGTERM Nov 27 05:10:33 localhost podman[326914]: 2025-11-27 10:10:33.455125179 +0000 UTC m=+0.067137248 container kill 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:10:33 localhost systemd[1]: libpod-72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd.scope: Deactivated successfully. Nov 27 05:10:33 localhost podman[326930]: 2025-11-27 10:10:33.533072538 +0000 UTC m=+0.052002830 container died 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.554 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:10:33 localhost systemd[1]: tmp-crun.qw7Caa.mount: Deactivated successfully. Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.556 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11134MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.556 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.557 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:10:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:33.560 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port f0b9f306-dd08-441a-8edb-4665d0586973 with type ""#033[00m Nov 27 05:10:33 localhost ovn_controller[156436]: 2025-11-27T10:10:33Z|00297|binding|INFO|Removing iface tap7223af45-8b ovn-installed in OVS Nov 27 05:10:33 localhost ovn_controller[156436]: 2025-11-27T10:10:33Z|00298|binding|INFO|Removing lport 7223af45-8bda-4c4b-93f0-864b601af7e7 ovn-installed in OVS Nov 27 05:10:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:33.563 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:0:ffff::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-12671eaa-864f-4aab-98f5-21123177cd60', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-12671eaa-864f-4aab-98f5-21123177cd60', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '292341d0c2ae492d9f1c680cc1b88f1e', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c207ff98-6fa5-4c37-ba07-efb5cc85a81c, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=7223af45-8bda-4c4b-93f0-864b601af7e7) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.564 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:33.566 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 7223af45-8bda-4c4b-93f0-864b601af7e7 in datapath 12671eaa-864f-4aab-98f5-21123177cd60 unbound from our chassis#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.571 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:33.572 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 12671eaa-864f-4aab-98f5-21123177cd60 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:10:33 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:33.574 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[1f8db6af-813f-437b-8011-43d13bbeedff]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:33 localhost podman[326930]: 2025-11-27 10:10:33.594952264 +0000 UTC m=+0.113882566 container remove 72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-12671eaa-864f-4aab-98f5-21123177cd60, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125) Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.605 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:33 localhost kernel: device tap7223af45-8b left promiscuous mode Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.628 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:33 localhost systemd[1]: libpod-conmon-72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd.scope: Deactivated successfully. Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.657 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.658 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.658 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:10:33 localhost nova_compute[284026]: 2025-11-27 10:10:33.709 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:10:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:33.718 265123 INFO neutron.agent.dhcp.agent [None req-aef86ec6-28b4-43a9-b02e-f6838e779dde - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:33 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:33.913 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:10:34 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/764883870' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:10:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:34.202 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '20'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.215 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.507s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.221 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.241 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.242 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.242 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.685s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:10:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:10:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:10:34 localhost systemd[1]: var-lib-containers-storage-overlay-6f81a373d2019bdd63d37cb55863549270978e47b342b29b79eecc8cd4f5fc1e-merged.mount: Deactivated successfully. Nov 27 05:10:34 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-72f7098d0e63d588cdc776d3813ac7ea69c6b101cb71bf25e60b695d24b2b7dd-userdata-shm.mount: Deactivated successfully. Nov 27 05:10:34 localhost systemd[1]: run-netns-qdhcp\x2d12671eaa\x2d864f\x2d4aab\x2d98f5\x2d21123177cd60.mount: Deactivated successfully. Nov 27 05:10:34 localhost podman[326976]: 2025-11-27 10:10:34.489671096 +0000 UTC m=+0.086066099 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible) Nov 27 05:10:34 localhost podman[326976]: 2025-11-27 10:10:34.505009389 +0000 UTC m=+0.101404382 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:10:34 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:10:34 localhost ovn_controller[156436]: 2025-11-27T10:10:34Z|00299|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:10:34 localhost nova_compute[284026]: 2025-11-27 10:10:34.612 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:34 localhost systemd[1]: tmp-crun.FvoHsC.mount: Deactivated successfully. Nov 27 05:10:34 localhost podman[326977]: 2025-11-27 10:10:34.633201221 +0000 UTC m=+0.226393768 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_id=multipathd, container_name=multipathd) Nov 27 05:10:34 localhost podman[326977]: 2025-11-27 10:10:34.646196231 +0000 UTC m=+0.239388778 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 05:10:34 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:10:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v450: 177 pgs: 177 active+clean; 146 MiB data, 900 MiB used, 41 GiB / 42 GiB avail; 18 KiB/s rd, 15 KiB/s wr, 25 op/s Nov 27 05:10:35 localhost nova_compute[284026]: 2025-11-27 10:10:35.773 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.242 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.243 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.243 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.366 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.367 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.367 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:10:36 localhost nova_compute[284026]: 2025-11-27 10:10:36.367 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:10:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:36.847 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:03:b1:49 10.100.0.18 10.100.0.3'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.18/28 10.100.0.3/28', 'neutron:device_id': 'ovnmeta-2d964b61-4b58-45cb-afaf-4a65506339b0', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-2d964b61-4b58-45cb-afaf-4a65506339b0', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=4a0fa5d2-996c-493e-8c9b-70b6dc7926b5, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=e3a3b0d7-be1b-4148-8ed4-3f81ff36e924) old=Port_Binding(mac=['fa:16:3e:03:b1:49 10.100.0.3'], external_ids={'neutron:cidrs': '10.100.0.3/28', 'neutron:device_id': 'ovnmeta-2d964b61-4b58-45cb-afaf-4a65506339b0', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-2d964b61-4b58-45cb-afaf-4a65506339b0', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '2', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:36.849 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port e3a3b0d7-be1b-4148-8ed4-3f81ff36e924 in datapath 2d964b61-4b58-45cb-afaf-4a65506339b0 updated#033[00m Nov 27 05:10:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:36.851 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 2d964b61-4b58-45cb-afaf-4a65506339b0, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:10:36 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:36.853 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e03632aa-0f3f-4647-8294-abd9515af92f]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:37 localhost nova_compute[284026]: 2025-11-27 10:10:37.105 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:37 localhost nova_compute[284026]: 2025-11-27 10:10:37.206 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:10:37 localhost nova_compute[284026]: 2025-11-27 10:10:37.241 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:10:37 localhost nova_compute[284026]: 2025-11-27 10:10:37.242 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:10:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v451: 177 pgs: 177 active+clean; 217 MiB data, 967 MiB used, 41 GiB / 42 GiB avail; 2.2 MiB/s rd, 3.0 MiB/s wr, 98 op/s Nov 27 05:10:37 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:10:37 localhost podman[327030]: 2025-11-27 10:10:37.592580104 +0000 UTC m=+0.065980907 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:10:37 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:10:37 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:10:37 localhost ovn_controller[156436]: 2025-11-27T10:10:37Z|00300|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:10:37 localhost nova_compute[284026]: 2025-11-27 10:10:37.923 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:38 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:38.093 2 INFO neutron.agent.securitygroups_rpc [None req-a2ded201-f513-4474-b855-295d5cc8959b 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:10:38 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/924381827' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:10:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:10:38 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/924381827' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:10:38 localhost podman[242678]: time="2025-11-27T10:10:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:10:38 localhost podman[242678]: @ - - [27/Nov/2025:10:10:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:10:38 localhost podman[242678]: @ - - [27/Nov/2025:10:10:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19744 "" "Go-http-client/1.1" Nov 27 05:10:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v452: 177 pgs: 177 active+clean; 217 MiB data, 967 MiB used, 41 GiB / 42 GiB avail; 2.1 MiB/s rd, 2.9 MiB/s wr, 94 op/s Nov 27 05:10:39 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:39.454 2 INFO neutron.agent.securitygroups_rpc [None req-2cf2fbe4-d1eb-4d88-84a4-66f75f08db42 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:40.294 2 INFO neutron.agent.securitygroups_rpc [None req-13307a95-6c10-432b-bf07-3856e585911e 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:40 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:40.364 2 INFO neutron.agent.securitygroups_rpc [None req-fbaf4aab-1e67-40c3-bee1-15600db4f4ae 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:40 localhost nova_compute[284026]: 2025-11-27 10:10:40.807 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:41 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:41.163 2 INFO neutron.agent.securitygroups_rpc [None req-d63dd633-4eb6-46c8-a330-058964f2acf7 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v453: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 3.1 MiB/s rd, 4.3 MiB/s wr, 148 op/s Nov 27 05:10:42 localhost nova_compute[284026]: 2025-11-27 10:10:42.133 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:42.414 2 INFO neutron.agent.securitygroups_rpc [None req-2b088175-b203-4652-989d-b2e3cb15c8f9 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:42 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:42.588 2 INFO neutron.agent.securitygroups_rpc [None req-2b088175-b203-4652-989d-b2e3cb15c8f9 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:10:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:10:42 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:10:42 localhost podman[327051]: 2025-11-27 10:10:42.991131326 +0000 UTC m=+0.082190864 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:10:43 localhost systemd[1]: tmp-crun.4YlsLd.mount: Deactivated successfully. Nov 27 05:10:43 localhost podman[327052]: 2025-11-27 10:10:43.055736376 +0000 UTC m=+0.142101227 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:10:43 localhost podman[327052]: 2025-11-27 10:10:43.06702549 +0000 UTC m=+0.153390381 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:10:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:43 localhost podman[327051]: 2025-11-27 10:10:43.077912952 +0000 UTC m=+0.168972490 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, container_name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:10:43 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:10:43 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:10:43 localhost podman[327053]: 2025-11-27 10:10:43.153284302 +0000 UTC m=+0.236221361 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, container_name=openstack_network_exporter, vendor=Red Hat, Inc., managed_by=edpm_ansible, release=1755695350, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, architecture=x86_64, distribution-scope=public, url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.tags=minimal rhel9, version=9.6, vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.buildah.version=1.33.7, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., maintainer=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, config_id=edpm) Nov 27 05:10:43 localhost podman[327053]: 2025-11-27 10:10:43.16581675 +0000 UTC m=+0.248753799 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-type=git, version=9.6, name=ubi9-minimal, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, maintainer=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, config_id=edpm, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, release=1755695350, io.openshift.tags=minimal rhel9, distribution-scope=public) Nov 27 05:10:43 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:10:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v454: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 2.9 MiB/s rd, 4.0 MiB/s wr, 139 op/s Nov 27 05:10:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:43.578 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:10:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:43.579 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:10:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:43.580 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:10:43 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:43.872 2 INFO neutron.agent.securitygroups_rpc [None req-76abc883-c0e7-42a4-873b-a3edc8a5a139 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:44.444 2 INFO neutron.agent.securitygroups_rpc [None req-2bff4705-ed6d-4ddb-942c-14d7f1e88438 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:10:44 Nov 27 05:10:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:10:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:10:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'manila_metadata', 'vms', '.mgr', 'volumes', 'manila_data', 'images'] Nov 27 05:10:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', ), ('cephfs', )] Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:10:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.004820107481853713 of space, bias 1.0, pg target 0.9640214963707425 quantized to 32 (current 32) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 2.453674623115578e-06 of space, bias 1.0, pg target 0.0004890991415410385 quantized to 32 (current 32) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8570103846780196 quantized to 32 (current 32) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.416259538432906e-05 quantized to 32 (current 32) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:10:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 6.979341150195422e-05 of space, bias 4.0, pg target 0.055462497673552956 quantized to 16 (current 16) Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:10:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:10:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v455: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 2.6 MiB/s rd, 3.6 MiB/s wr, 123 op/s Nov 27 05:10:45 localhost nova_compute[284026]: 2025-11-27 10:10:45.810 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #40. Immutable memtables: 0. Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.883255) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 21] Flushing memtable with next log file: 40 Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238245883313, "job": 21, "event": "flush_started", "num_memtables": 1, "num_entries": 2102, "num_deletes": 272, "total_data_size": 3614851, "memory_usage": 3732848, "flush_reason": "Manual Compaction"} Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 21] Level-0 flush table #41: started Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238245904237, "cf_name": "default", "job": 21, "event": "table_file_creation", "file_number": 41, "file_size": 2352768, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 24043, "largest_seqno": 26140, "table_properties": {"data_size": 2344260, "index_size": 5141, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2309, "raw_key_size": 20489, "raw_average_key_size": 22, "raw_value_size": 2326347, "raw_average_value_size": 2548, "num_data_blocks": 215, "num_entries": 913, "num_filter_entries": 913, "num_deletions": 272, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238147, "oldest_key_time": 1764238147, "file_creation_time": 1764238245, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 41, "seqno_to_time_mapping": "N/A"}} Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 21] Flush lasted 21051 microseconds, and 6793 cpu microseconds. Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.904299) [db/flush_job.cc:967] [default] [JOB 21] Level-0 flush table #41: 2352768 bytes OK Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.904331) [db/memtable_list.cc:519] [default] Level-0 commit table #41 started Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.908481) [db/memtable_list.cc:722] [default] Level-0 commit table #41: memtable #1 done Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.908509) EVENT_LOG_v1 {"time_micros": 1764238245908502, "job": 21, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.908561) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 21] Try to delete WAL files size 3604881, prev total WAL file size 3605601, number of live WAL files 2. Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000037.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.915516) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003132323939' seq:72057594037927935, type:22 .. '7061786F73003132353531' seq:0, type:0; will stop at (end) Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 22] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 21 Base level 0, inputs: [41(2297KB)], [39(17MB)] Nov 27 05:10:45 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238245915624, "job": 22, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [41], "files_L6": [39], "score": -1, "input_data_size": 20575894, "oldest_snapshot_seqno": -1} Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 22] Generated table #42: 13018 keys, 19288924 bytes, temperature: kUnknown Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238246047473, "cf_name": "default", "job": 22, "event": "table_file_creation", "file_number": 42, "file_size": 19288924, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19215433, "index_size": 39854, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 32581, "raw_key_size": 349312, "raw_average_key_size": 26, "raw_value_size": 18994564, "raw_average_value_size": 1459, "num_data_blocks": 1496, "num_entries": 13018, "num_filter_entries": 13018, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238245, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 42, "seqno_to_time_mapping": "N/A"}} Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.048107) [db/compaction/compaction_job.cc:1663] [default] [JOB 22] Compacted 1@0 + 1@6 files to L6 => 19288924 bytes Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.050198) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 155.8 rd, 146.0 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(2.2, 17.4 +0.0 blob) out(18.4 +0.0 blob), read-write-amplify(16.9) write-amplify(8.2) OK, records in: 13577, records dropped: 559 output_compression: NoCompression Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.050234) EVENT_LOG_v1 {"time_micros": 1764238246050217, "job": 22, "event": "compaction_finished", "compaction_time_micros": 132108, "compaction_time_cpu_micros": 53990, "output_level": 6, "num_output_files": 1, "total_output_size": 19288924, "num_input_records": 13577, "num_output_records": 13018, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000041.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238246050981, "job": 22, "event": "table_file_deletion", "file_number": 41} Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000039.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238246054855, "job": 22, "event": "table_file_deletion", "file_number": 39} Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:45.915334) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.055046) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.055055) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.055058) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.055061) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:10:46.055063) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:10:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:46.517 265123 INFO neutron.agent.linux.ip_lib [None req-89e1fdee-beb1-47c0-a606-eb4dac81ed68 - - - - - -] Device tape40f0d62-66 cannot be used as it has no MAC address#033[00m Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.548 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost kernel: device tape40f0d62-66 entered promiscuous mode Nov 27 05:10:46 localhost NetworkManager[5971]: [1764238246.5617] manager: (tape40f0d62-66): new Generic device (/org/freedesktop/NetworkManager/Devices/51) Nov 27 05:10:46 localhost ovn_controller[156436]: 2025-11-27T10:10:46Z|00301|binding|INFO|Claiming lport e40f0d62-660f-418c-81b2-7fd6d49626dc for this chassis. Nov 27 05:10:46 localhost ovn_controller[156436]: 2025-11-27T10:10:46Z|00302|binding|INFO|e40f0d62-660f-418c-81b2-7fd6d49626dc: Claiming unknown Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.562 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost systemd-udevd[327121]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:10:46 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:46.573 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=2b858f17-8826-413f-b25c-b932d021210b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e40f0d62-660f-418c-81b2-7fd6d49626dc) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:46 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:46.575 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e40f0d62-660f-418c-81b2-7fd6d49626dc in datapath b7d49570-5e12-4b69-a59c-f0ae4ffaef4b bound to our chassis#033[00m Nov 27 05:10:46 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:46.576 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network b7d49570-5e12-4b69-a59c-f0ae4ffaef4b or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:10:46 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:46.577 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[e075ed87-307d-4d7a-b316-b52548bfeb97]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.600 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost ovn_controller[156436]: 2025-11-27T10:10:46Z|00303|binding|INFO|Setting lport e40f0d62-660f-418c-81b2-7fd6d49626dc ovn-installed in OVS Nov 27 05:10:46 localhost ovn_controller[156436]: 2025-11-27T10:10:46Z|00304|binding|INFO|Setting lport e40f0d62-660f-418c-81b2-7fd6d49626dc up in Southbound Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.606 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.634 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost nova_compute[284026]: 2025-11-27 10:10:46.658 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:46.933 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:10:46Z, description=, device_id=4399e1d6-206a-46f3-be12-26760c2b4d66, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=51404f3c-e737-496b-b077-ea70dacdc908, ip_allocation=immediate, mac_address=fa:16:3e:ef:47:e9, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=2788, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:10:46Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:10:47 localhost nova_compute[284026]: 2025-11-27 10:10:47.160 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:47 localhost podman[327168]: 2025-11-27 10:10:47.16234651 +0000 UTC m=+0.075481273 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:10:47 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:10:47 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:10:47 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:10:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/.meta.tmp' Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/.meta.tmp' to config b'/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/.meta' Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "format": "json"}]: dispatch Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v456: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 3.7 MiB/s rd, 3.6 MiB/s wr, 160 op/s Nov 27 05:10:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:47.402 265123 INFO neutron.agent.dhcp.agent [None req-6aba26e4-8f7d-44eb-95d0-1443336e3ca3 - - - - - -] DHCP configuration for ports {'51404f3c-e737-496b-b077-ea70dacdc908'} is completed#033[00m Nov 27 05:10:47 localhost podman[327215]: Nov 27 05:10:47 localhost podman[327215]: 2025-11-27 10:10:47.574741805 +0000 UTC m=+0.099721277 container create 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:10:47 localhost systemd[1]: Started libpod-conmon-2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf.scope. Nov 27 05:10:47 localhost podman[327215]: 2025-11-27 10:10:47.526620169 +0000 UTC m=+0.051599661 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:10:47 localhost systemd[1]: Started libcrun container. Nov 27 05:10:47 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/439fb133029af46d2ccaad4b4d4c2a360fb192869e7bc5667620768360744f31/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:10:47 localhost podman[327215]: 2025-11-27 10:10:47.649264691 +0000 UTC m=+0.174244163 container init 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:10:47 localhost podman[327215]: 2025-11-27 10:10:47.658356467 +0000 UTC m=+0.183335929 container start 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:10:47 localhost dnsmasq[327233]: started, version 2.85 cachesize 150 Nov 27 05:10:47 localhost dnsmasq[327233]: DNS service limited to local subnets Nov 27 05:10:47 localhost dnsmasq[327233]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:10:47 localhost dnsmasq[327233]: warning: no upstream servers configured Nov 27 05:10:47 localhost dnsmasq-dhcp[327233]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:10:47 localhost dnsmasq[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/addn_hosts - 0 addresses Nov 27 05:10:47 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/host Nov 27 05:10:47 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/opts Nov 27 05:10:47 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:47.743 2 INFO neutron.agent.securitygroups_rpc [None req-3000a4ad-8ad3-47eb-a9b4-f4d74a302e87 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:47 localhost nova_compute[284026]: 2025-11-27 10:10:47.853 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:47.861 265123 INFO neutron.agent.dhcp.agent [None req-deb17666-e4d6-4270-8b9b-f14221c50028 - - - - - -] DHCP configuration for ports {'7883436a-9b2f-421d-bdd9-3c6b8eaac66f'} is completed#033[00m Nov 27 05:10:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:48 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:48.430 2 INFO neutron.agent.securitygroups_rpc [None req-fb50a792-6884-4341-bead-0e4aea013ea8 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:48.551 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:10:47Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=8542686f-90bf-46c3-9dc2-bd24c84a88c3, ip_allocation=immediate, mac_address=fa:16:3e:c3:2c:2a, name=tempest-PortsTestJSON-1325560547, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:10:44Z, description=, dns_domain=, id=b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-PortsTestJSON-1575576927, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=13607, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=2779, status=ACTIVE, subnets=['98122da9-f9ad-4bdf-a9f1-e4ee5bf53c1b'], tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:10:45Z, vlan_transparent=None, network_id=b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4588f774-ca3d-4240-83a8-bd827b113565'], standard_attr_id=2797, status=DOWN, tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:10:47Z on network b7d49570-5e12-4b69-a59c-f0ae4ffaef4b#033[00m Nov 27 05:10:48 localhost dnsmasq[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/addn_hosts - 1 addresses Nov 27 05:10:48 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/host Nov 27 05:10:48 localhost podman[327250]: 2025-11-27 10:10:48.883494954 +0000 UTC m=+0.062674209 container kill 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:10:48 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/opts Nov 27 05:10:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:49.044 2 INFO neutron.agent.securitygroups_rpc [None req-782a2600-1ebf-414e-b5a9-c7e7def35b59 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:49 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:49.126 265123 INFO neutron.agent.dhcp.agent [None req-4c43855f-acfa-42fd-9f36-bba8e909165e - - - - - -] DHCP configuration for ports {'8542686f-90bf-46c3-9dc2-bd24c84a88c3'} is completed#033[00m Nov 27 05:10:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v457: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 1.1 MiB/s wr, 97 op/s Nov 27 05:10:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:49.707 2 INFO neutron.agent.securitygroups_rpc [None req-a1c1ba13-efa7-43ce-ac6c-6f64198cbad6 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:49 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:49.769 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:10:49Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=55292ce4-ce58-4083-a5cf-4e5b394ead26, ip_allocation=immediate, mac_address=fa:16:3e:44:14:0f, name=tempest-PortsTestJSON-980567381, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:10:44Z, description=, dns_domain=, id=b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-PortsTestJSON-1575576927, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=13607, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=2779, status=ACTIVE, subnets=['98122da9-f9ad-4bdf-a9f1-e4ee5bf53c1b'], tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:10:45Z, vlan_transparent=None, network_id=b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4588f774-ca3d-4240-83a8-bd827b113565'], standard_attr_id=2799, status=DOWN, tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:10:49Z on network b7d49570-5e12-4b69-a59c-f0ae4ffaef4b#033[00m Nov 27 05:10:50 localhost systemd[1]: tmp-crun.7JuFHY.mount: Deactivated successfully. Nov 27 05:10:50 localhost dnsmasq[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/addn_hosts - 2 addresses Nov 27 05:10:50 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/host Nov 27 05:10:50 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/opts Nov 27 05:10:50 localhost podman[327290]: 2025-11-27 10:10:50.018109105 +0000 UTC m=+0.075972797 container kill 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:10:50 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:50.292 265123 INFO neutron.agent.dhcp.agent [None req-c16df722-0110-4bc8-903e-f03b925e0fba - - - - - -] DHCP configuration for ports {'55292ce4-ce58-4083-a5cf-4e5b394ead26'} is completed#033[00m Nov 27 05:10:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "auth_id": "tempest-cephx-id-338434859", "tenant_id": "d64da1b3f8234f34bd513ae567e381c2", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:10:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume authorize, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, tenant_id:d64da1b3f8234f34bd513ae567e381c2, vol_name:cephfs) < "" Nov 27 05:10:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} v 0) Nov 27 05:10:50 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} : dispatch Nov 27 05:10:50 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-338434859 with tenant d64da1b3f8234f34bd513ae567e381c2 Nov 27 05:10:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-338434859", "caps": ["mds", "allow rw path=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666", "osd", "allow rw pool=manila_data namespace=fsvolumens_5b66d55f-fbd5-4547-90d2-df58efa8d841", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:10:50 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-338434859", "caps": ["mds", "allow rw path=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666", "osd", "allow rw pool=manila_data namespace=fsvolumens_5b66d55f-fbd5-4547-90d2-df58efa8d841", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:10:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume authorize, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, tenant_id:d64da1b3f8234f34bd513ae567e381c2, vol_name:cephfs) < "" Nov 27 05:10:50 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:50.674 2 INFO neutron.agent.securitygroups_rpc [None req-1e2091ac-4e0c-4e4f-8e8d-45c7d0a1a2b7 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:50 localhost nova_compute[284026]: 2025-11-27 10:10:50.813 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:50 localhost dnsmasq[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/addn_hosts - 1 addresses Nov 27 05:10:50 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/host Nov 27 05:10:50 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/opts Nov 27 05:10:50 localhost podman[327328]: 2025-11-27 10:10:50.910275697 +0000 UTC m=+0.059082852 container kill 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:10:51 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} : dispatch Nov 27 05:10:51 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-338434859", "caps": ["mds", "allow rw path=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666", "osd", "allow rw pool=manila_data namespace=fsvolumens_5b66d55f-fbd5-4547-90d2-df58efa8d841", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:10:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-338434859", "caps": ["mds", "allow rw path=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666", "osd", "allow rw pool=manila_data namespace=fsvolumens_5b66d55f-fbd5-4547-90d2-df58efa8d841", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:10:51 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-338434859", "caps": ["mds", "allow rw path=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666", "osd", "allow rw pool=manila_data namespace=fsvolumens_5b66d55f-fbd5-4547-90d2-df58efa8d841", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:10:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v458: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 1.9 MiB/s rd, 1.1 MiB/s wr, 112 op/s Nov 27 05:10:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "auth_id": "tempest-cephx-id-338434859", "format": "json"}]: dispatch Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume deauthorize, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} v 0) Nov 27 05:10:51 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} : dispatch Nov 27 05:10:51 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-338434859"} v 0) Nov 27 05:10:51 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-338434859"} : dispatch Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume deauthorize, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "auth_id": "tempest-cephx-id-338434859", "format": "json"}]: dispatch Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume evict, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-338434859, client_metadata.root=/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841/7e95d905-a1ea-4ec8-a288-df5a39528666 Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-338434859, format:json, prefix:fs subvolume evict, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:51 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:51.626 2 INFO neutron.agent.securitygroups_rpc [None req-d6d775ba-acc3-46dd-b9d5-14e558abbdea 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:10:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "format": "json"}]: dispatch Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:10:51 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5b66d55f-fbd5-4547-90d2-df58efa8d841' of type subvolume Nov 27 05:10:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:10:51.885+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5b66d55f-fbd5-4547-90d2-df58efa8d841' of type subvolume Nov 27 05:10:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "5b66d55f-fbd5-4547-90d2-df58efa8d841", "force": true, "format": "json"}]: dispatch Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:10:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/5b66d55f-fbd5-4547-90d2-df58efa8d841'' moved to trashcan Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:10:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5b66d55f-fbd5-4547-90d2-df58efa8d841, vol_name:cephfs) < "" Nov 27 05:10:52 localhost podman[327364]: 2025-11-27 10:10:52.006807663 +0000 UTC m=+0.100037385 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_id=ovn_controller, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:10:52 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-338434859", "format": "json"} : dispatch Nov 27 05:10:52 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-338434859"} : dispatch Nov 27 05:10:52 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-338434859"} : dispatch Nov 27 05:10:52 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-338434859"}]': finished Nov 27 05:10:52 localhost dnsmasq[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/addn_hosts - 0 addresses Nov 27 05:10:52 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/host Nov 27 05:10:52 localhost dnsmasq-dhcp[327233]: read /var/lib/neutron/dhcp/b7d49570-5e12-4b69-a59c-f0ae4ffaef4b/opts Nov 27 05:10:52 localhost podman[327381]: 2025-11-27 10:10:52.02861412 +0000 UTC m=+0.080360825 container kill 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 05:10:52 localhost podman[327364]: 2025-11-27 10:10:52.06464319 +0000 UTC m=+0.157872962 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:10:52 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:10:52 localhost podman[327365]: 2025-11-27 10:10:52.074030293 +0000 UTC m=+0.161115180 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:10:52 localhost podman[327365]: 2025-11-27 10:10:52.154916781 +0000 UTC m=+0.242001668 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:10:52 localhost nova_compute[284026]: 2025-11-27 10:10:52.163 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:52 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:10:52 localhost dnsmasq[327233]: exiting on receipt of SIGTERM Nov 27 05:10:52 localhost podman[327448]: 2025-11-27 10:10:52.847678674 +0000 UTC m=+0.063890191 container kill 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:10:52 localhost systemd[1]: libpod-2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf.scope: Deactivated successfully. Nov 27 05:10:52 localhost podman[327460]: 2025-11-27 10:10:52.924079091 +0000 UTC m=+0.061359663 container died 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:10:52 localhost podman[327460]: 2025-11-27 10:10:52.960979505 +0000 UTC m=+0.098260037 container cleanup 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:10:52 localhost systemd[1]: libpod-conmon-2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf.scope: Deactivated successfully. Nov 27 05:10:52 localhost systemd[1]: var-lib-containers-storage-overlay-439fb133029af46d2ccaad4b4d4c2a360fb192869e7bc5667620768360744f31-merged.mount: Deactivated successfully. Nov 27 05:10:52 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf-userdata-shm.mount: Deactivated successfully. Nov 27 05:10:53 localhost podman[327462]: 2025-11-27 10:10:53.005504974 +0000 UTC m=+0.133962759 container remove 2911e15034542886bcfdc217f99a4d5afed6a56e5a35e09cfbf2531edad6ebcf (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3) Nov 27 05:10:53 localhost nova_compute[284026]: 2025-11-27 10:10:53.017 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:53 localhost ovn_controller[156436]: 2025-11-27T10:10:53Z|00305|binding|INFO|Releasing lport e40f0d62-660f-418c-81b2-7fd6d49626dc from this chassis (sb_readonly=0) Nov 27 05:10:53 localhost ovn_controller[156436]: 2025-11-27T10:10:53Z|00306|binding|INFO|Setting lport e40f0d62-660f-418c-81b2-7fd6d49626dc down in Southbound Nov 27 05:10:53 localhost kernel: device tape40f0d62-66 left promiscuous mode Nov 27 05:10:53 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:53.029 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-b7d49570-5e12-4b69-a59c-f0ae4ffaef4b', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=2b858f17-8826-413f-b25c-b932d021210b, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=e40f0d62-660f-418c-81b2-7fd6d49626dc) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:53 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:53.031 162092 INFO neutron.agent.ovn.metadata.agent [-] Port e40f0d62-660f-418c-81b2-7fd6d49626dc in datapath b7d49570-5e12-4b69-a59c-f0ae4ffaef4b unbound from our chassis#033[00m Nov 27 05:10:53 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:53.034 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network b7d49570-5e12-4b69-a59c-f0ae4ffaef4b, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:10:53 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:53.035 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[ff2932f3-ba89-41b4-87fc-c472da944366]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:53 localhost nova_compute[284026]: 2025-11-27 10:10:53.048 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:53 localhost nova_compute[284026]: 2025-11-27 10:10:53.266 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v459: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 1.1 MiB/s rd, 8.7 KiB/s wr, 51 op/s Nov 27 05:10:53 localhost systemd[1]: run-netns-qdhcp\x2db7d49570\x2d5e12\x2d4b69\x2da59c\x2df0ae4ffaef4b.mount: Deactivated successfully. Nov 27 05:10:53 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:53.421 265123 INFO neutron.agent.dhcp.agent [None req-6b7e173d-6abe-4659-b864-d7017ed8164a - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:53 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:53.422 265123 INFO neutron.agent.dhcp.agent [None req-6b7e173d-6abe-4659-b864-d7017ed8164a - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:53 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:53.553 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:10:54 localhost ovn_controller[156436]: 2025-11-27T10:10:54Z|00307|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:10:54 localhost nova_compute[284026]: 2025-11-27 10:10:54.190 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v460: 177 pgs: 177 active+clean; 193 MiB data, 982 MiB used, 41 GiB / 42 GiB avail; 1.1 MiB/s rd, 8.7 KiB/s wr, 51 op/s Nov 27 05:10:55 localhost openstack_network_exporter[244641]: ERROR 10:10:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:10:55 localhost openstack_network_exporter[244641]: ERROR 10:10:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:10:55 localhost openstack_network_exporter[244641]: ERROR 10:10:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:10:55 localhost openstack_network_exporter[244641]: ERROR 10:10:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:10:55 localhost openstack_network_exporter[244641]: Nov 27 05:10:55 localhost openstack_network_exporter[244641]: ERROR 10:10:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:10:55 localhost openstack_network_exporter[244641]: Nov 27 05:10:55 localhost nova_compute[284026]: 2025-11-27 10:10:55.814 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:57 localhost nova_compute[284026]: 2025-11-27 10:10:57.168 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v461: 177 pgs: 177 active+clean; 226 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 3.1 MiB/s rd, 2.2 MiB/s wr, 126 op/s Nov 27 05:10:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e210 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:10:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:58.298 265123 INFO neutron.agent.linux.ip_lib [None req-c6b472ea-de90-4580-bcaa-ecd22552a683 - - - - - -] Device tapcf9be4ab-fb cannot be used as it has no MAC address#033[00m Nov 27 05:10:58 localhost nova_compute[284026]: 2025-11-27 10:10:58.322 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:58 localhost kernel: device tapcf9be4ab-fb entered promiscuous mode Nov 27 05:10:58 localhost NetworkManager[5971]: [1764238258.3301] manager: (tapcf9be4ab-fb): new Generic device (/org/freedesktop/NetworkManager/Devices/52) Nov 27 05:10:58 localhost nova_compute[284026]: 2025-11-27 10:10:58.330 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:58 localhost ovn_controller[156436]: 2025-11-27T10:10:58Z|00308|binding|INFO|Claiming lport cf9be4ab-fb7e-4560-8278-5e4ae34881c0 for this chassis. Nov 27 05:10:58 localhost ovn_controller[156436]: 2025-11-27T10:10:58Z|00309|binding|INFO|cf9be4ab-fb7e-4560-8278-5e4ae34881c0: Claiming unknown Nov 27 05:10:58 localhost systemd-udevd[327502]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:10:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:58.341 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=7f521b81-63fa-4e4b-8271-ef478d7f47a5, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=cf9be4ab-fb7e-4560-8278-5e4ae34881c0) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:10:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:58.343 162092 INFO neutron.agent.ovn.metadata.agent [-] Port cf9be4ab-fb7e-4560-8278-5e4ae34881c0 in datapath 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9 bound to our chassis#033[00m Nov 27 05:10:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:58.344 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:10:58 localhost ovn_metadata_agent[162087]: 2025-11-27 10:10:58.345 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3487d66c-04a1-46a3-80f6-774b035d1abf]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost ovn_controller[156436]: 2025-11-27T10:10:58Z|00310|binding|INFO|Setting lport cf9be4ab-fb7e-4560-8278-5e4ae34881c0 ovn-installed in OVS Nov 27 05:10:58 localhost ovn_controller[156436]: 2025-11-27T10:10:58Z|00311|binding|INFO|Setting lport cf9be4ab-fb7e-4560-8278-5e4ae34881c0 up in Southbound Nov 27 05:10:58 localhost nova_compute[284026]: 2025-11-27 10:10:58.362 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost journal[232028]: ethtool ioctl error on tapcf9be4ab-fb: No such device Nov 27 05:10:58 localhost nova_compute[284026]: 2025-11-27 10:10:58.412 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:58 localhost nova_compute[284026]: 2025-11-27 10:10:58.448 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:10:58 localhost neutron_sriov_agent[258162]: 2025-11-27 10:10:58.868 2 INFO neutron.agent.securitygroups_rpc [None req-5135f863-ae50-4dea-acf8-938e7f8b6d04 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:10:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v462: 177 pgs: 177 active+clean; 226 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 2.0 MiB/s rd, 2.2 MiB/s wr, 89 op/s Nov 27 05:10:59 localhost podman[327573]: Nov 27 05:10:59 localhost podman[327573]: 2025-11-27 10:10:59.547011501 +0000 UTC m=+0.090429995 container create 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:10:59 localhost systemd[1]: Started libpod-conmon-047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5.scope. Nov 27 05:10:59 localhost systemd[1]: tmp-crun.epHrhL.mount: Deactivated successfully. Nov 27 05:10:59 localhost podman[327573]: 2025-11-27 10:10:59.503403338 +0000 UTC m=+0.046821862 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:10:59 localhost systemd[1]: Started libcrun container. Nov 27 05:10:59 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/4706baa38edebc6bfb83900043ca48303cb1d5c66477bf929e622e3b6c2355de/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:10:59 localhost podman[327573]: 2025-11-27 10:10:59.626480742 +0000 UTC m=+0.169899246 container init 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) Nov 27 05:10:59 localhost podman[327573]: 2025-11-27 10:10:59.635471624 +0000 UTC m=+0.178890118 container start 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3) Nov 27 05:10:59 localhost dnsmasq[327591]: started, version 2.85 cachesize 150 Nov 27 05:10:59 localhost dnsmasq[327591]: DNS service limited to local subnets Nov 27 05:10:59 localhost dnsmasq[327591]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:10:59 localhost dnsmasq[327591]: warning: no upstream servers configured Nov 27 05:10:59 localhost dnsmasq-dhcp[327591]: DHCP, static leases only on 10.100.0.0, lease time 1d Nov 27 05:10:59 localhost dnsmasq[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/addn_hosts - 0 addresses Nov 27 05:10:59 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/host Nov 27 05:10:59 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/opts Nov 27 05:10:59 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:10:59.829 265123 INFO neutron.agent.dhcp.agent [None req-bdd466ab-ce0b-423c-81dc-b86faa23f309 - - - - - -] DHCP configuration for ports {'6f9c1e79-8394-4dcb-be6e-ae2f9c8975ee'} is completed#033[00m Nov 27 05:11:00 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:00.262 2 INFO neutron.agent.securitygroups_rpc [None req-9a24e805-2c2f-473f-9c89-0f7cb6494297 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e211 e211: 6 total, 6 up, 6 in Nov 27 05:11:00 localhost nova_compute[284026]: 2025-11-27 10:11:00.818 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:00 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:00.964 2 INFO neutron.agent.securitygroups_rpc [None req-1be774a1-73f4-4146-8947-80a602757669 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:01 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:11:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:11:01 localhost podman[327609]: 2025-11-27 10:11:01.050628878 +0000 UTC m=+0.065785712 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:11:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:11:01 localhost ovn_controller[156436]: 2025-11-27T10:11:01Z|00312|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:11:01 localhost nova_compute[284026]: 2025-11-27 10:11:01.207 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v464: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 2.5 MiB/s rd, 4.7 MiB/s wr, 135 op/s Nov 27 05:11:02 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:02.002 2 INFO neutron.agent.securitygroups_rpc [None req-89d70196-6ffa-4b1b-a750-b8706bbe3192 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:02 localhost nova_compute[284026]: 2025-11-27 10:11:02.171 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:03.072 2 INFO neutron.agent.securitygroups_rpc [None req-6db47004-e89a-402f-8458-ea63e8dd4c36 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:11:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e211 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:03 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:03.169 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:11:01Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=ec187c7b-7ca0-4917-b7fd-af4211e2123c, ip_allocation=immediate, mac_address=fa:16:3e:a3:2d:86, name=tempest-PortsTestJSON-2000438837, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:10:54Z, description=, dns_domain=, id=5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-PortsTestJSON-1287112975, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=46346, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=2821, status=ACTIVE, subnets=['05ef9a3c-0d4e-482d-bc29-81cb72660891'], tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:10:57Z, vlan_transparent=None, network_id=5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['4588f774-ca3d-4240-83a8-bd827b113565'], standard_attr_id=2849, status=DOWN, tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:11:01Z on network 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9#033[00m Nov 27 05:11:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v465: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 2.5 MiB/s rd, 4.7 MiB/s wr, 135 op/s Nov 27 05:11:03 localhost dnsmasq[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/addn_hosts - 1 addresses Nov 27 05:11:03 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/host Nov 27 05:11:03 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/opts Nov 27 05:11:03 localhost podman[327648]: 2025-11-27 10:11:03.436499341 +0000 UTC m=+0.070900541 container kill 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:11:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/.meta.tmp' Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/.meta.tmp' to config b'/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/.meta' Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "format": "json"}]: dispatch Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:03.662 2 INFO neutron.agent.securitygroups_rpc [None req-2ab58a80-a9e3-49ff-8b77-5cc87ace7c6f 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:03 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:03.705 265123 INFO neutron.agent.dhcp.agent [None req-7aab93ee-284c-4f0c-a987-e94cb3f1e7f6 - - - - - -] DHCP configuration for ports {'ec187c7b-7ca0-4917-b7fd-af4211e2123c'} is completed#033[00m Nov 27 05:11:04 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:04.176 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=21, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=20) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:04 localhost nova_compute[284026]: 2025-11-27 10:11:04.177 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:04 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:04.178 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 7 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:11:04 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:04.645 2 INFO neutron.agent.securitygroups_rpc [None req-ffceec98-b822-4bea-a5a5-b68242f98a04 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:11:04 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:11:04 localhost podman[327668]: 2025-11-27 10:11:04.988039069 +0000 UTC m=+0.080934111 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:11:05 localhost podman[327668]: 2025-11-27 10:11:05.000484643 +0000 UTC m=+0.093379695 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:11:05 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:11:05 localhost podman[327669]: 2025-11-27 10:11:05.089245464 +0000 UTC m=+0.180825971 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, config_id=multipathd, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:11:05 localhost podman[327669]: 2025-11-27 10:11:05.100290631 +0000 UTC m=+0.191871118 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible) Nov 27 05:11:05 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:11:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v466: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 2.5 MiB/s rd, 4.7 MiB/s wr, 135 op/s Nov 27 05:11:05 localhost nova_compute[284026]: 2025-11-27 10:11:05.852 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:06 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:06.066 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:11:01Z, description=, device_id=92311736-0f51-44e4-9800-006dd3ab1956, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=ec187c7b-7ca0-4917-b7fd-af4211e2123c, ip_allocation=immediate, mac_address=fa:16:3e:a3:2d:86, name=tempest-PortsTestJSON-2000438837, network_id=5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, port_security_enabled=True, project_id=0de604c307c6461fbd4a8776c4a2a1fa, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=3, security_groups=['4588f774-ca3d-4240-83a8-bd827b113565'], standard_attr_id=2849, status=ACTIVE, tags=[], tenant_id=0de604c307c6461fbd4a8776c4a2a1fa, updated_at=2025-11-27T10:11:04Z on network 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9#033[00m Nov 27 05:11:06 localhost dnsmasq[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/addn_hosts - 1 addresses Nov 27 05:11:06 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/host Nov 27 05:11:06 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/opts Nov 27 05:11:06 localhost podman[327722]: 2025-11-27 10:11:06.304115254 +0000 UTC m=+0.064209428 container kill 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:11:06 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:06.600 265123 INFO neutron.agent.dhcp.agent [None req-065b5b5e-5193-4b33-aabf-588040b7a111 - - - - - -] DHCP configuration for ports {'ec187c7b-7ca0-4917-b7fd-af4211e2123c'} is completed#033[00m Nov 27 05:11:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve49", "tenant_id": "378926434bb940509dd59eb2cb5a02f9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:11:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve49, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve49", "format": "json"} v 0) Nov 27 05:11:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve49", "format": "json"} : dispatch Nov 27 05:11:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID eve49 with tenant 378926434bb940509dd59eb2cb5a02f9 Nov 27 05:11:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.eve49", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:11:07 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve49", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve49, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:07 localhost nova_compute[284026]: 2025-11-27 10:11:07.175 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v467: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 65 KiB/s rd, 2.1 MiB/s wr, 93 op/s Nov 27 05:11:07 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:07.593 2 INFO neutron.agent.securitygroups_rpc [None req-879ddb37-2e37-4ff6-a921-4c9e4899d008 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:11:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:11:07 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:11:07 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:11:07 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev b6cf1a71-653c-46e0-9344-8417ef3a7a2f (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:11:07 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev b6cf1a71-653c-46e0-9344-8417ef3a7a2f (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:11:07 localhost ceph-mgr[290377]: [progress INFO root] Completed event b6cf1a71-653c-46e0-9344-8417ef3a7a2f (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:11:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:11:07 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:11:07 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:07.840 2 INFO neutron.agent.securitygroups_rpc [None req-f5144b44-3de5-4b62-b3b4-014e18c0efb6 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:07 localhost dnsmasq[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/addn_hosts - 0 addresses Nov 27 05:11:07 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/host Nov 27 05:11:07 localhost podman[327829]: 2025-11-27 10:11:07.852070115 +0000 UTC m=+0.062100603 container kill 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:11:07 localhost dnsmasq-dhcp[327591]: read /var/lib/neutron/dhcp/5f320ba4-28a2-4c00-afd0-82ffd91a4fc9/opts Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve49", "format": "json"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve49", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve49", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.eve49", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:11:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:11:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e212 e212: 6 total, 6 up, 6 in Nov 27 05:11:08 localhost nova_compute[284026]: 2025-11-27 10:11:08.070 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:08 localhost kernel: device tapcf9be4ab-fb left promiscuous mode Nov 27 05:11:08 localhost ovn_controller[156436]: 2025-11-27T10:11:08Z|00313|binding|INFO|Releasing lport cf9be4ab-fb7e-4560-8278-5e4ae34881c0 from this chassis (sb_readonly=0) Nov 27 05:11:08 localhost ovn_controller[156436]: 2025-11-27T10:11:08Z|00314|binding|INFO|Setting lport cf9be4ab-fb7e-4560-8278-5e4ae34881c0 down in Southbound Nov 27 05:11:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e212 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:08.089 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=7f521b81-63fa-4e4b-8271-ef478d7f47a5, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=cf9be4ab-fb7e-4560-8278-5e4ae34881c0) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:08.092 162092 INFO neutron.agent.ovn.metadata.agent [-] Port cf9be4ab-fb7e-4560-8278-5e4ae34881c0 in datapath 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9 unbound from our chassis#033[00m Nov 27 05:11:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:08.094 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:11:08 localhost nova_compute[284026]: 2025-11-27 10:11:08.095 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:08 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:08.095 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[8a3085b5-d72f-4ea5-a336-e158e8f97c1d]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:11:08 localhost podman[242678]: time="2025-11-27T10:11:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:11:08 localhost podman[242678]: @ - - [27/Nov/2025:10:11:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159336 "" "Go-http-client/1.1" Nov 27 05:11:08 localhost podman[242678]: @ - - [27/Nov/2025:10:11:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20225 "" "Go-http-client/1.1" Nov 27 05:11:08 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:08.920 2 INFO neutron.agent.securitygroups_rpc [None req-feeec61a-b172-414d-a8a8-29c05ce4bf24 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:09 localhost dnsmasq[327591]: exiting on receipt of SIGTERM Nov 27 05:11:09 localhost podman[327886]: 2025-11-27 10:11:09.114381485 +0000 UTC m=+0.063497382 container kill 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:11:09 localhost systemd[1]: libpod-047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5.scope: Deactivated successfully. Nov 27 05:11:09 localhost podman[327901]: 2025-11-27 10:11:09.190428932 +0000 UTC m=+0.059184054 container died 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:11:09 localhost podman[327901]: 2025-11-27 10:11:09.22972406 +0000 UTC m=+0.098479142 container cleanup 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:11:09 localhost systemd[1]: libpod-conmon-047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5.scope: Deactivated successfully. Nov 27 05:11:09 localhost podman[327902]: 2025-11-27 10:11:09.271191677 +0000 UTC m=+0.135484120 container remove 047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-5f320ba4-28a2-4c00-afd0-82ffd91a4fc9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:11:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:09.299 265123 INFO neutron.agent.dhcp.agent [None req-a90a9f9f-b34b-479d-8bd1-3864f4741a71 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:11:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v469: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 43 KiB/s rd, 6.2 KiB/s wr, 57 op/s Nov 27 05:11:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:09.393 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:11:09 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:09.625 2 INFO neutron.agent.securitygroups_rpc [None req-6c9a7af7-3169-43c3-933d-c0a5ad0adb80 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:09 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:11:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:11:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:11:09 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3280515883' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:11:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:11:09 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3280515883' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:11:10 localhost systemd[1]: var-lib-containers-storage-overlay-4706baa38edebc6bfb83900043ca48303cb1d5c66477bf929e622e3b6c2355de-merged.mount: Deactivated successfully. Nov 27 05:11:10 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-047693e5aa76297bfeb8690a09a7be4353f20b97f174b3052b98acaae1401ca5-userdata-shm.mount: Deactivated successfully. Nov 27 05:11:10 localhost systemd[1]: run-netns-qdhcp\x2d5f320ba4\x2d28a2\x2d4c00\x2dafd0\x2d82ffd91a4fc9.mount: Deactivated successfully. Nov 27 05:11:10 localhost ovn_controller[156436]: 2025-11-27T10:11:10Z|00315|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:11:10 localhost nova_compute[284026]: 2025-11-27 10:11:10.215 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve48", "tenant_id": "378926434bb940509dd59eb2cb5a02f9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:11:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve48, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve48", "format": "json"} v 0) Nov 27 05:11:10 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve48", "format": "json"} : dispatch Nov 27 05:11:10 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID eve48 with tenant 378926434bb940509dd59eb2cb5a02f9 Nov 27 05:11:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.eve48", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:11:10 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve48", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve48, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:10 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:10.453 2 INFO neutron.agent.securitygroups_rpc [None req-5649097a-407d-4db4-9bad-ef92cf756031 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:11:10 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve48", "format": "json"} : dispatch Nov 27 05:11:10 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve48", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve48", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.eve48", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:11:10 localhost nova_compute[284026]: 2025-11-27 10:11:10.896 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:11.179 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '21'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:11:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v470: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 26 KiB/s wr, 107 op/s Nov 27 05:11:12 localhost nova_compute[284026]: 2025-11-27 10:11:12.181 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:12 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:12.671 2 INFO neutron.agent.securitygroups_rpc [None req-a3080488-50ff-4ca0-95e7-b23a7870f3dc 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['aa905413-f97c-4ada-890b-760520a5d954']#033[00m Nov 27 05:11:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e212 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v471: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 26 KiB/s wr, 107 op/s Nov 27 05:11:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:11:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:11:13 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:11:13 localhost systemd[1]: tmp-crun.CvldC6.mount: Deactivated successfully. Nov 27 05:11:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve48", "format": "json"}]: dispatch Nov 27 05:11:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:eve48, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:14 localhost systemd[1]: tmp-crun.cXsJOG.mount: Deactivated successfully. Nov 27 05:11:14 localhost podman[327929]: 2025-11-27 10:11:14.060964276 +0000 UTC m=+0.152465196 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.schema-version=1.0, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=edpm) Nov 27 05:11:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve48", "format": "json"} v 0) Nov 27 05:11:14 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve48", "format": "json"} : dispatch Nov 27 05:11:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.eve48"} v 0) Nov 27 05:11:14 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve48"} : dispatch Nov 27 05:11:14 localhost podman[327931]: 2025-11-27 10:11:14.018206105 +0000 UTC m=+0.102485500 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, io.buildah.version=1.33.7, release=1755695350, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, vcs-type=git, distribution-scope=public, build-date=2025-08-20T13:12:41, config_id=edpm, io.openshift.expose-services=, name=ubi9-minimal, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64) Nov 27 05:11:14 localhost podman[327930]: 2025-11-27 10:11:13.985878895 +0000 UTC m=+0.077424566 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:11:14 localhost podman[327931]: 2025-11-27 10:11:14.099686039 +0000 UTC m=+0.183965454 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm, name=ubi9-minimal, io.openshift.tags=minimal rhel9, vcs-type=git, version=9.6, container_name=openstack_network_exporter, io.buildah.version=1.33.7, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, maintainer=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, vendor=Red Hat, Inc.) Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:eve48, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:14 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:11:14 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve48", "format": "json"}]: dispatch Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:eve48, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:14 localhost podman[327930]: 2025-11-27 10:11:14.115781083 +0000 UTC m=+0.207326714 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=eve48, client_metadata.root=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:eve48, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:14 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:11:14 localhost podman[327929]: 2025-11-27 10:11:14.150857197 +0000 UTC m=+0.242358047 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3) Nov 27 05:11:14 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:14 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:14.576 2 INFO neutron.agent.securitygroups_rpc [None req-ffec6525-5082-4333-be49-3c0305f69f9e 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve48", "format": "json"} : dispatch Nov 27 05:11:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve48"} : dispatch Nov 27 05:11:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve48"} : dispatch Nov 27 05:11:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.eve48"}]': finished Nov 27 05:11:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:14.955 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:67:78:8e 10.100.0.18 10.100.0.2'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.18/28 10.100.0.2/28', 'neutron:device_id': 'ovnmeta-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=fb3e362b-46dc-4de9-b3e9-3c6d7f94c653, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=e5ae7735-3041-4b5f-b21f-fc7cca474260) old=Port_Binding(mac=['fa:16:3e:67:78:8e 10.100.0.2'], external_ids={'neutron:cidrs': '10.100.0.2/28', 'neutron:device_id': 'ovnmeta-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '2', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:14.958 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port e5ae7735-3041-4b5f-b21f-fc7cca474260 in datapath caac68ed-83b8-4412-9ac0-b1c245b30c7f updated#033[00m Nov 27 05:11:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:14.960 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network caac68ed-83b8-4412-9ac0-b1c245b30c7f, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:11:14 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:14.961 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[b799b9d3-a2ae-4b0b-ba3f-9b612e222285]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:11:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:15.127 2 INFO neutron.agent.securitygroups_rpc [None req-5a4b15ce-5914-41d0-b63a-a383313e00d1 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v472: 177 pgs: 177 active+clean; 272 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 79 KiB/s rd, 26 KiB/s wr, 107 op/s Nov 27 05:11:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:15.639 2 INFO neutron.agent.securitygroups_rpc [None req-af524eef-a19d-4aec-87f6-9ed9d6754e10 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['aa905413-f97c-4ada-890b-760520a5d954', '22987901-cd29-471d-949f-c3646dc0dc90']#033[00m Nov 27 05:11:15 localhost nova_compute[284026]: 2025-11-27 10:11:15.933 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:15 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:15.989 2 INFO neutron.agent.securitygroups_rpc [None req-6523776b-1a62-4d00-9ee9-246914da069a 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:16.195 2 INFO neutron.agent.securitygroups_rpc [None req-e3191dd5-d2e7-428d-89a4-3b16dd90a176 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['22987901-cd29-471d-949f-c3646dc0dc90']#033[00m Nov 27 05:11:16 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:16.326 2 INFO neutron.agent.securitygroups_rpc [None req-57c2193b-c243-49fb-904b-268d9d663807 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:17 localhost nova_compute[284026]: 2025-11-27 10:11:17.211 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v473: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 47 KiB/s rd, 32 KiB/s wr, 67 op/s Nov 27 05:11:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve47", "tenant_id": "378926434bb940509dd59eb2cb5a02f9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:11:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve47, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve47", "format": "json"} v 0) Nov 27 05:11:17 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve47", "format": "json"} : dispatch Nov 27 05:11:17 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID eve47 with tenant 378926434bb940509dd59eb2cb5a02f9 Nov 27 05:11:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.eve47", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:11:17 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve47", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:eve47, format:json, prefix:fs subvolume authorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, tenant_id:378926434bb940509dd59eb2cb5a02f9, vol_name:cephfs) < "" Nov 27 05:11:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve47", "format": "json"} : dispatch Nov 27 05:11:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve47", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.eve47", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.eve47", "caps": ["mds", "allow rw path=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d", "osd", "allow rw pool=manila_data namespace=fsvolumens_b96eb12a-0de9-479f-848a-c5e98607b8dc", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:11:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e213 e213: 6 total, 6 up, 6 in Nov 27 05:11:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e213 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:18 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "effef14e-3ae5-4220-ac9c-9ab067db38e8", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:18 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e214 e214: 6 total, 6 up, 6 in Nov 27 05:11:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/effef14e-3ae5-4220-ac9c-9ab067db38e8/.meta.tmp' Nov 27 05:11:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/effef14e-3ae5-4220-ac9c-9ab067db38e8/.meta.tmp' to config b'/volumes/_nogroup/effef14e-3ae5-4220-ac9c-9ab067db38e8/.meta' Nov 27 05:11:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "effef14e-3ae5-4220-ac9c-9ab067db38e8", "format": "json"}]: dispatch Nov 27 05:11:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:19 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:19.159 2 INFO neutron.agent.securitygroups_rpc [None req-89309257-5349-43c1-a5f0-93154d1d0205 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v476: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 6.1 KiB/s rd, 14 KiB/s wr, 10 op/s Nov 27 05:11:19 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:19.488 2 INFO neutron.agent.securitygroups_rpc [None req-1fa1fb2a-2957-457e-80f5-9aaf14241c52 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['e45e435a-53a3-458a-ad18-d95618050ae0']#033[00m Nov 27 05:11:20 localhost nova_compute[284026]: 2025-11-27 10:11:20.959 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:21 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve47", "format": "json"}]: dispatch Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:eve47, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve47", "format": "json"} v 0) Nov 27 05:11:21 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve47", "format": "json"} : dispatch Nov 27 05:11:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.eve47"} v 0) Nov 27 05:11:21 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve47"} : dispatch Nov 27 05:11:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v477: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 32 KiB/s rd, 35 KiB/s wr, 52 op/s Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:eve47, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:21 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve47", "format": "json"}]: dispatch Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:eve47, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=eve47, client_metadata.root=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:11:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:eve47, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:21 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:21.783 2 INFO neutron.agent.securitygroups_rpc [None req-a83b2ce4-69a3-4123-b450-6da3bbb219a6 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:21.856 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['fa:16:3e:67:78:8e 10.100.0.18 10.100.0.2 10.100.0.34'], port_security=[], type=localport, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': ''}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.100.0.18/28 10.100.0.2/28 10.100.0.34/28', 'neutron:device_id': 'ovnmeta-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '6', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=fb3e362b-46dc-4de9-b3e9-3c6d7f94c653, chassis=[], tunnel_key=1, gateway_chassis=[], requested_chassis=[], logical_port=e5ae7735-3041-4b5f-b21f-fc7cca474260) old=Port_Binding(mac=['fa:16:3e:67:78:8e 10.100.0.18 10.100.0.2'], external_ids={'neutron:cidrs': '10.100.0.18/28 10.100.0.2/28', 'neutron:device_id': 'ovnmeta-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:device_owner': 'network:distributed', 'neutron:mtu': '', 'neutron:network_name': 'neutron-caac68ed-83b8-4412-9ac0-b1c245b30c7f', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '0de604c307c6461fbd4a8776c4a2a1fa', 'neutron:revision_number': '5', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:21.858 162092 INFO neutron.agent.ovn.metadata.agent [-] Metadata Port e5ae7735-3041-4b5f-b21f-fc7cca474260 in datapath caac68ed-83b8-4412-9ac0-b1c245b30c7f updated#033[00m Nov 27 05:11:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:21.861 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network caac68ed-83b8-4412-9ac0-b1c245b30c7f, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:11:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:21.862 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[0883da99-999b-442b-b387-28b1d82a1b53]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:11:22 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve47", "format": "json"} : dispatch Nov 27 05:11:22 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve47"} : dispatch Nov 27 05:11:22 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve47"} : dispatch Nov 27 05:11:22 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.eve47"}]': finished Nov 27 05:11:22 localhost nova_compute[284026]: 2025-11-27 10:11:22.239 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e215 e215: 6 total, 6 up, 6 in Nov 27 05:11:22 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:22.513 2 INFO neutron.agent.securitygroups_rpc [None req-c7a36cd4-2ab9-4f3c-ad08-5eb7f164063e 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['e45e435a-53a3-458a-ad18-d95618050ae0', '2f15744a-f468-4752-9468-cb9463da0d06', '1204a0c3-6abf-4ccf-a575-eabb48241915']#033[00m Nov 27 05:11:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:11:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:11:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "effef14e-3ae5-4220-ac9c-9ab067db38e8", "format": "json"}]: dispatch Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:23 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:23.020+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'effef14e-3ae5-4220-ac9c-9ab067db38e8' of type subvolume Nov 27 05:11:23 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'effef14e-3ae5-4220-ac9c-9ab067db38e8' of type subvolume Nov 27 05:11:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "effef14e-3ae5-4220-ac9c-9ab067db38e8", "force": true, "format": "json"}]: dispatch Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:23 localhost podman[327989]: 2025-11-27 10:11:23.027452318 +0000 UTC m=+0.108539394 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/effef14e-3ae5-4220-ac9c-9ab067db38e8'' moved to trashcan Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:effef14e-3ae5-4220-ac9c-9ab067db38e8, vol_name:cephfs) < "" Nov 27 05:11:23 localhost podman[327989]: 2025-11-27 10:11:23.068191285 +0000 UTC m=+0.149278331 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0) Nov 27 05:11:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e215 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:23 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:11:23 localhost podman[327990]: 2025-11-27 10:11:23.114021649 +0000 UTC m=+0.192532785 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:11:23 localhost podman[327990]: 2025-11-27 10:11:23.130064831 +0000 UTC m=+0.208575967 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:11:23 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:23.134 2 INFO neutron.agent.securitygroups_rpc [None req-5c8abfe8-fcfe-4b55-828e-84e3113319b3 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['2f15744a-f468-4752-9468-cb9463da0d06', '1204a0c3-6abf-4ccf-a575-eabb48241915']#033[00m Nov 27 05:11:23 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:11:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v479: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 35 KiB/s rd, 29 KiB/s wr, 54 op/s Nov 27 05:11:23 localhost nova_compute[284026]: 2025-11-27 10:11:23.740 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e216 e216: 6 total, 6 up, 6 in Nov 27 05:11:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve49", "format": "json"}]: dispatch Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:eve49, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.eve49", "format": "json"} v 0) Nov 27 05:11:25 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve49", "format": "json"} : dispatch Nov 27 05:11:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.eve49"} v 0) Nov 27 05:11:25 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve49"} : dispatch Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:eve49, format:json, prefix:fs subvolume deauthorize, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "auth_id": "eve49", "format": "json"}]: dispatch Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:eve49, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=eve49, client_metadata.root=/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc/6744e541-7a82-47c1-84d7-3e66ff69968d Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:eve49, format:json, prefix:fs subvolume evict, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v481: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 27 KiB/s wr, 52 op/s Nov 27 05:11:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:25.371 2 INFO neutron.agent.securitygroups_rpc [None req-26f35802-f941-47a8-9e8b-f65619118ca1 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['24c412c6-96ad-405c-8a72-3cf4d486913f']#033[00m Nov 27 05:11:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "format": "json"}]: dispatch Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:25.442+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b96eb12a-0de9-479f-848a-c5e98607b8dc' of type subvolume Nov 27 05:11:25 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b96eb12a-0de9-479f-848a-c5e98607b8dc' of type subvolume Nov 27 05:11:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "b96eb12a-0de9-479f-848a-c5e98607b8dc", "force": true, "format": "json"}]: dispatch Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/b96eb12a-0de9-479f-848a-c5e98607b8dc'' moved to trashcan Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b96eb12a-0de9-479f-848a-c5e98607b8dc, vol_name:cephfs) < "" Nov 27 05:11:25 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:25.492 2 INFO neutron.agent.securitygroups_rpc [None req-4462b48e-f9bf-4210-bccb-9cda9b6632ee 9aabda1741b34698990e1de942126449 0de604c307c6461fbd4a8776c4a2a1fa - - default default] Security group member updated ['4588f774-ca3d-4240-83a8-bd827b113565']#033[00m Nov 27 05:11:25 localhost nova_compute[284026]: 2025-11-27 10:11:25.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:25 localhost openstack_network_exporter[244641]: ERROR 10:11:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:11:25 localhost openstack_network_exporter[244641]: ERROR 10:11:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:11:25 localhost openstack_network_exporter[244641]: ERROR 10:11:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:11:25 localhost openstack_network_exporter[244641]: ERROR 10:11:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:11:25 localhost openstack_network_exporter[244641]: Nov 27 05:11:25 localhost openstack_network_exporter[244641]: ERROR 10:11:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:11:25 localhost openstack_network_exporter[244641]: Nov 27 05:11:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e217 e217: 6 total, 6 up, 6 in Nov 27 05:11:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.eve49", "format": "json"} : dispatch Nov 27 05:11:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve49"} : dispatch Nov 27 05:11:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.eve49"} : dispatch Nov 27 05:11:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.eve49"}]': finished Nov 27 05:11:25 localhost nova_compute[284026]: 2025-11-27 10:11:25.961 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:26 localhost nova_compute[284026]: 2025-11-27 10:11:26.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:26 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:26.746 2 INFO neutron.agent.securitygroups_rpc [None req-8f781833-efb1-4370-8104-99eea2c8d39f 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['759ea3bb-7688-4988-ab6d-615a54ccb912', '24c412c6-96ad-405c-8a72-3cf4d486913f']#033[00m Nov 27 05:11:27 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:27.093 2 INFO neutron.agent.securitygroups_rpc [None req-64ac22b7-75e7-4bca-b9e6-e25a5a58a64c 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['759ea3bb-7688-4988-ab6d-615a54ccb912']#033[00m Nov 27 05:11:27 localhost nova_compute[284026]: 2025-11-27 10:11:27.279 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v483: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 89 KiB/s rd, 49 KiB/s wr, 126 op/s Nov 27 05:11:27 localhost nova_compute[284026]: 2025-11-27 10:11:27.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:27 localhost nova_compute[284026]: 2025-11-27 10:11:27.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:27 localhost nova_compute[284026]: 2025-11-27 10:11:27.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:11:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e217 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e218 e218: 6 total, 6 up, 6 in Nov 27 05:11:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:11:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "format": "json"}]: dispatch Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:11:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:11:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v485: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 89 KiB/s rd, 49 KiB/s wr, 126 op/s Nov 27 05:11:29 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:29.333 2 INFO neutron.agent.securitygroups_rpc [None req-79125b86-0049-4ea7-bf56-3b3528c71535 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['eb84378f-56f5-4089-acdb-8bac5bd0cb14']#033[00m Nov 27 05:11:29 localhost nova_compute[284026]: 2025-11-27 10:11:29.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e219 e219: 6 total, 6 up, 6 in Nov 27 05:11:30 localhost nova_compute[284026]: 2025-11-27 10:11:30.755 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:30 localhost nova_compute[284026]: 2025-11-27 10:11:30.994 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v487: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 126 KiB/s rd, 68 KiB/s wr, 179 op/s Nov 27 05:11:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:31.326 2 INFO neutron.agent.securitygroups_rpc [None req-c97c4512-4d2f-4ba8-a877-6ae7ba411dc7 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['eb84378f-56f5-4089-acdb-8bac5bd0cb14', 'be8cc068-3670-4c8b-94b3-eecaa629df62', '9ccf38c1-126b-4ff9-a8ab-4d7ade7551a9']#033[00m Nov 27 05:11:31 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:31.770 2 INFO neutron.agent.securitygroups_rpc [None req-51395775-cadc-4d0a-a4f2-e81813556348 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['9ccf38c1-126b-4ff9-a8ab-4d7ade7551a9', 'be8cc068-3670-4c8b-94b3-eecaa629df62']#033[00m Nov 27 05:11:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e220 e220: 6 total, 6 up, 6 in Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.315 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "44ede067-a156-4a97-8295-4fdfc8788b8e", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/44ede067-a156-4a97-8295-4fdfc8788b8e/.meta.tmp' Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/44ede067-a156-4a97-8295-4fdfc8788b8e/.meta.tmp' to config b'/volumes/_nogroup/44ede067-a156-4a97-8295-4fdfc8788b8e/.meta' Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "44ede067-a156-4a97-8295-4fdfc8788b8e", "format": "json"}]: dispatch Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e221 e221: 6 total, 6 up, 6 in Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.770 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.770 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.770 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.771 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:11:32 localhost nova_compute[284026]: 2025-11-27 10:11:32.771 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:11:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e221 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:11:33 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3349845793' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.233 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.462s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:11:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v490: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 50 KiB/s rd, 25 KiB/s wr, 70 op/s Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.411 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.411 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.640 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.642 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11114MB free_disk=41.70015335083008GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.642 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.643 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.788 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.788 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.789 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:11:33 localhost nova_compute[284026]: 2025-11-27 10:11:33.838 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:11:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e222 e222: 6 total, 6 up, 6 in Nov 27 05:11:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:11:34 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2028243713' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:11:34 localhost nova_compute[284026]: 2025-11-27 10:11:34.336 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.497s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:11:34 localhost nova_compute[284026]: 2025-11-27 10:11:34.343 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:11:34 localhost nova_compute[284026]: 2025-11-27 10:11:34.364 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:11:34 localhost nova_compute[284026]: 2025-11-27 10:11:34.368 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:11:34 localhost nova_compute[284026]: 2025-11-27 10:11:34.368 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.725s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:11:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/.meta.tmp' Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/.meta.tmp' to config b'/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/.meta' Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:11:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "format": "json"}]: dispatch Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:11:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:11:34 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:34.926 2 INFO neutron.agent.securitygroups_rpc [None req-2ec6f34b-acd5-4ef6-9500-1a5025adecd5 2c95dab969e440e18e033d18348fd311 513ae51ffd00467497896e0aa512bbc2 - - default default] Security group member updated ['552cd642-1baf-49a0-b7e4-19f7d263c456']#033[00m Nov 27 05:11:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v492: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 41 KiB/s rd, 21 KiB/s wr, 58 op/s Nov 27 05:11:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume resize", "vol_name": "cephfs", "sub_name": "44ede067-a156-4a97-8295-4fdfc8788b8e", "new_size": 2147483648, "format": "json"}]: dispatch Nov 27 05:11:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:11:35 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:11:36 localhost nova_compute[284026]: 2025-11-27 10:11:36.041 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:36 localhost podman[328085]: 2025-11-27 10:11:36.043628365 +0000 UTC m=+0.132950422 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, container_name=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, tcib_managed=true, config_id=multipathd, managed_by=edpm_ansible, org.label-schema.schema-version=1.0) Nov 27 05:11:36 localhost podman[328085]: 2025-11-27 10:11:36.051512857 +0000 UTC m=+0.140834914 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:11:36 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:11:36 localhost podman[328084]: 2025-11-27 10:11:36.063459109 +0000 UTC m=+0.154094821 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:11:36 localhost podman[328084]: 2025-11-27 10:11:36.098959652 +0000 UTC m=+0.189595394 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent) Nov 27 05:11:36 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:11:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e/.meta.tmp' Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e/.meta.tmp' to config b'/volumes/_nogroup/ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e/.meta' Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:11:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e", "format": "json"}]: dispatch Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:11:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.317 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v493: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 40 KiB/s rd, 22 KiB/s wr, 57 op/s Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.369 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.369 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.370 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.480 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.481 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.481 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:11:37 localhost nova_compute[284026]: 2025-11-27 10:11:37.482 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:11:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e222 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:38 localhost podman[242678]: time="2025-11-27T10:11:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:11:38 localhost podman[242678]: @ - - [27/Nov/2025:10:11:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:11:38 localhost podman[242678]: @ - - [27/Nov/2025:10:11:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19751 "" "Go-http-client/1.1" Nov 27 05:11:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "44ede067-a156-4a97-8295-4fdfc8788b8e", "format": "json"}]: dispatch Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:44ede067-a156-4a97-8295-4fdfc8788b8e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:44ede067-a156-4a97-8295-4fdfc8788b8e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:38 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:38.895+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '44ede067-a156-4a97-8295-4fdfc8788b8e' of type subvolume Nov 27 05:11:38 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '44ede067-a156-4a97-8295-4fdfc8788b8e' of type subvolume Nov 27 05:11:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "44ede067-a156-4a97-8295-4fdfc8788b8e", "force": true, "format": "json"}]: dispatch Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/44ede067-a156-4a97-8295-4fdfc8788b8e'' moved to trashcan Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:44ede067-a156-4a97-8295-4fdfc8788b8e, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:39 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:39.174 2 INFO neutron.agent.securitygroups_rpc [req-353cd1bb-5b03-4267-9b07-99acb9b31e1a req-e7c8c992-96ee-4178-98d9-d577de427baf c86b8f3bf6f9490288bf21849ce8e6b5 67fb135d336b4706a482e380561998d6 - - default default] Security group member updated ['8afab950-f78e-4af9-961b-2530a2994b49']#033[00m Nov 27 05:11:39 localhost nova_compute[284026]: 2025-11-27 10:11:39.177 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:11:39 localhost nova_compute[284026]: 2025-11-27 10:11:39.195 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:11:39 localhost nova_compute[284026]: 2025-11-27 10:11:39.195 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/.meta.tmp' Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/.meta.tmp' to config b'/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/.meta' Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "format": "json"}]: dispatch Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v494: 177 pgs: 177 active+clean; 273 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 32 KiB/s rd, 18 KiB/s wr, 46 op/s Nov 27 05:11:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "ec2c5e75-9fb0-4002-ac50-e25621663620", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ec2c5e75-9fb0-4002-ac50-e25621663620/.meta.tmp' Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ec2c5e75-9fb0-4002-ac50-e25621663620/.meta.tmp' to config b'/volumes/_nogroup/ec2c5e75-9fb0-4002-ac50-e25621663620/.meta' Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "ec2c5e75-9fb0-4002-ac50-e25621663620", "format": "json"}]: dispatch Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:41 localhost nova_compute[284026]: 2025-11-27 10:11:41.067 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v495: 177 pgs: 177 active+clean; 194 MiB data, 984 MiB used, 41 GiB / 42 GiB avail; 54 KiB/s rd, 41 KiB/s wr, 82 op/s Nov 27 05:11:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "7260cd56-3d12-425d-bc6a-99686549f079", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/7260cd56-3d12-425d-bc6a-99686549f079/.meta.tmp' Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/7260cd56-3d12-425d-bc6a-99686549f079/.meta.tmp' to config b'/volumes/_nogroup/7260cd56-3d12-425d-bc6a-99686549f079/.meta' Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "7260cd56-3d12-425d-bc6a-99686549f079", "format": "json"}]: dispatch Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:42 localhost nova_compute[284026]: 2025-11-27 10:11:42.320 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:11:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:11:42 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:42 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:11:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b", "osd", "allow rw pool=manila_data namespace=fsvolumens_ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:11:42 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b", "osd", "allow rw pool=manila_data namespace=fsvolumens_ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:11:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e223 e223: 6 total, 6 up, 6 in Nov 27 05:11:42 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:42 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b", "osd", "allow rw pool=manila_data namespace=fsvolumens_ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:42 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b", "osd", "allow rw pool=manila_data namespace=fsvolumens_ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:42 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b", "osd", "allow rw pool=manila_data namespace=fsvolumens_ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:11:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e223 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v497: 177 pgs: 177 active+clean; 194 MiB data, 984 MiB used, 41 GiB / 42 GiB avail; 50 KiB/s rd, 38 KiB/s wr, 75 op/s Nov 27 05:11:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:43.580 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:11:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:43.580 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:11:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:11:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e224 e224: 6 total, 6 up, 6 in Nov 27 05:11:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "ec2c5e75-9fb0-4002-ac50-e25621663620", "format": "json"}]: dispatch Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:ec2c5e75-9fb0-4002-ac50-e25621663620, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:ec2c5e75-9fb0-4002-ac50-e25621663620, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:43 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:43.844+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ec2c5e75-9fb0-4002-ac50-e25621663620' of type subvolume Nov 27 05:11:43 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ec2c5e75-9fb0-4002-ac50-e25621663620' of type subvolume Nov 27 05:11:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "ec2c5e75-9fb0-4002-ac50-e25621663620", "force": true, "format": "json"}]: dispatch Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/ec2c5e75-9fb0-4002-ac50-e25621663620'' moved to trashcan Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ec2c5e75-9fb0-4002-ac50-e25621663620, vol_name:cephfs) < "" Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.142 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.174 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.174 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c6d7a4ad-d989-4312-bc8f-b9e7fe5b8170', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.143973', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79b313ba-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '11feb2a5e661ce1c68f0fa608853d9bec4e7059a922b26d4930856ab22b5af6d'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.143973', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79b328c8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '41ad26555a7613cbd63157340d7496c98136e9ea6c85fac77f6c5a9df3b40a85'}]}, 'timestamp': '2025-11-27 10:11:44.175345', '_unique_id': 'f541cd79f5df405abc7b7723ba2cf3ee'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.177 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.178 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.182 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '006bf496-0503-48bb-b85d-1e5764113e46', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.178634', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79b448b6-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': 'f407bea62193643959025ef85f518c8d4c8a9f6a5f2a9cb5607d6a84d4c1bf2a'}]}, 'timestamp': '2025-11-27 10:11:44.182769', '_unique_id': '1d46e2432a184647bf07f983fe502473'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.183 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.185 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.195 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.195 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '69d0fac1-b544-4156-805a-52eb24f6cac0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.185176', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79b64e68-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': 'c203e9df56d2ce35d823d217d6a3c3f14ab470903e70529f31b2e07841808f9e'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.185176', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79b65f98-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': '278da14453f3aa6c6938652e981833cba51139243d935cb9155161d5d3a32560'}]}, 'timestamp': '2025-11-27 10:11:44.196422', '_unique_id': '4e2ac29e199d4132919df1b7feed3a30'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.197 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.198 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.214 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b85e2e9d-1db5-4133-95e6-fc652312c3df', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:11:44.198721', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '79b94136-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.402757721, 'message_signature': '2af44bb358209fc702bcaafa13be7e62385c46fb206f44ac4697990e0ac95fcd'}]}, 'timestamp': '2025-11-27 10:11:44.215293', '_unique_id': '71fc76efd23546569cce959780ef6d7e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.216 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.217 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '04a3bdcf-91b1-474c-a1a7-e3d523d242c3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.217543', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79b9ab76-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '8e420f7fdbcb7a0c8b0f644423d73ab12980435bb75028da4457447f42106e84'}]}, 'timestamp': '2025-11-27 10:11:44.218030', '_unique_id': 'd76f8784f2f94e11937d0edf29e5e46f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.218 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.220 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.221 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.221 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9f65165b-c3ae-48dd-bf93-a55446f3fa11', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.221098', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79ba3820-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '91cc0a20b1cd9d25240a3bb2cc5ea59580b1518bdde02a40840457614a748108'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.221098', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79ba6214-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '5cd782c675ab02ae8d6049402de15b0258162e2c19dfce4a1c40f3fc455ec6e9'}]}, 'timestamp': '2025-11-27 10:11:44.222721', '_unique_id': 'af7a429a111a4833b1ded042ae8498f5'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.223 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.225 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.225 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ab414260-6dc6-46a9-ae4f-fbcfba17ae3c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.226144', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79bafe7c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '8563f48d44450f03dd0ea2473a1120aa0df8fd569029c8b9a6a05fd178bf5a24'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.226144', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79bb2596-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '5b7947599808e69dbd1cd77b2031f3c4118446c6ad4874612e31b29359008a7f'}]}, 'timestamp': '2025-11-27 10:11:44.227777', '_unique_id': '1d64e7fc2f8341a0b1bd52e972b70fb9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9d632ab1-7338-4e3b-8917-c1759dc5f408', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.233382', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79bc1ea6-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '1ad99aa9cb0bede25ff8d481449172a8e2d1e6dcdcb599a441ed9cddf38d6311'}]}, 'timestamp': '2025-11-27 10:11:44.235300', '_unique_id': 'a5c49559922a4555bcb1d6c07e50a7d1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.240 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.240 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.241 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '281c5829-33f0-4de4-9559-f8ac3242d9f0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.240439', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79bd32dc-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': 'e955c03d6f56dc10da4dced3cf511131aaeed2ceb7c55be139288da2fe026086'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.240439', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79bd45c4-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': '40e2c133dea107db4266ae4e117f9498cb1c8656d87c02efe87668500993de88'}]}, 'timestamp': '2025-11-27 10:11:44.241725', '_unique_id': '053365fe010e401e99a4da3e79b3258c'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.242 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.243 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'f9f331e2-c334-4819-b4ff-2befd133d816', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.244113', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79bdb89c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '7ffe1bf5615daa15da924ac571575f62acd5b8dd635aaff4289d50eab0f22fe5'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.244113', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79bdcba2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': 'cc0ae3e4c9e3a88e57af16d49a695f6cfcb03344141ca2bc5783773b88fd988a'}]}, 'timestamp': '2025-11-27 10:11:44.245036', '_unique_id': 'fd5e8208bcec4614807171b3ad06b518'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.245 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0a7ce4d7-f243-41fe-8328-8bc5210df116', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.247246', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79be32c2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': 'a9763ca7653308f42c35782a66e77afe6485d3208f5c729a9461537d68d4350a'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.247246', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79be4474-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '0e00489cc097fb1c1fde42eb8cb2bbf09ee5a38c7517c4eb435c1ca24fffad33'}]}, 'timestamp': '2025-11-27 10:11:44.248126', '_unique_id': '3c0b0e80d3c245049a232dcc1aed7082'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a8d5e26c-dbfd-4680-978c-15ef59b32a13', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.250514', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79beb5b2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': 'f5416df1ea9aacea5f46b8ead615f9a72a29efd50e4fa386ad6efc9ac622b7e0'}]}, 'timestamp': '2025-11-27 10:11:44.251083', '_unique_id': 'c12d50d0dbd847e4bbd4569235eb22b3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '6f66eecb-c630-4d02-8e26-0acb5eaf5399', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.253468', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79bf27e0-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '6c9ad60d6b07a5b8ecc1149203b278d45b73e5f92194592d82244e90855f6caf'}]}, 'timestamp': '2025-11-27 10:11:44.253980', '_unique_id': '61e0f0ebc6e44fad975d722e40a13692'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '38829877-da0b-4c5c-aab9-6c39bc56a5a3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.256161', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79bf90ae-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '237d5bd07a47b6425aa9fe6a4e28684b71ead0e24c386b63b296c45946dd5647'}]}, 'timestamp': '2025-11-27 10:11:44.256701', '_unique_id': '353653702fbe48b1ad826fd6b9ea39be'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.258 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.258 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7567b204-d275-45cb-9252-52718d41abf1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.258109', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79bfd80c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': 'd576e62771400bacc6365706e3a85ad9d3dcb2d7f960f77537b2ee78e4913fed'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.258109', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79bfe342-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.373234589, 'message_signature': '3c8791637d98339ec990b052bfeedc6695c7c9d1496f6f3a7ed678c7fe6c459f'}]}, 'timestamp': '2025-11-27 10:11:44.258674', '_unique_id': '3cc39aea4cd74346ad4d63035603b8d0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.259 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.260 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.260 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 18020000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '59d12d5e-8684-4a5e-85f3-6491cfeedf86', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 18020000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:11:44.260142', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '79c02758-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.402757721, 'message_signature': '592aeff54f8b06c20731ef7b5226f50b63adcf0c3948d2bd1c8f6136d0925af7'}]}, 'timestamp': '2025-11-27 10:11:44.260440', '_unique_id': '3061523ce81748e295857a960c66ce84'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ac3c0fe8-c9f2-4e74-bb57-7eab7c08eb4f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.261871', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79c06baa-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '8b9c2831ce4d913703c4905915cf091d954b214a0b8efbce8215545bb6ac633d'}]}, 'timestamp': '2025-11-27 10:11:44.262190', '_unique_id': '63f2522da452443281de7cb832e268b2'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.263 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.263 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '700b27a2-7d54-46fd-aadf-00955695e755', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.264007', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79c0c046-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '38aeca529759c89af3d939094cec833bbdd0daaf4dd6bb47e62adefeb2d6b79e'}]}, 'timestamp': '2025-11-27 10:11:44.264347', '_unique_id': '4ecfd621a80944b6bccdfa2428976c5a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.265 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.266 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.266 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a5ceca4b-bcb1-411b-b6eb-5cc51c352fd6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.266273', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79c117da-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '8385c3c84d700ef2438f2f1c46a0df65089d22605a1a302d4eb0a0ee4705ee4a'}]}, 'timestamp': '2025-11-27 10:11:44.266705', '_unique_id': 'c671aedad61b42d08069e730fad1177f'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.267 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.268 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.268 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7ea6fa29-a8b8-445d-9fdd-64b556b98d7f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:11:44.268335', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '79c16a5a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.366696403, 'message_signature': '9c99731be6f41da1e4b154a55a280e3f83665a722b20706fd7db2b70ad796647'}]}, 'timestamp': '2025-11-27 10:11:44.268701', '_unique_id': '10105aba8cd84b079f2435c646ff1148'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.269 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.270 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.270 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.270 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.270 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3c14a78b-4e3d-459f-9692-e2381f901dfa', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:11:44.270455', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '79c1bae6-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': '4db9c6f28442a8f14b6e2c84ee528dabdac54b8cf9385d16d9bc04127ba3ac44'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:11:44.270455', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '79c1c522-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12534.332036882, 'message_signature': 'bcc515a39882e34b8dd279a5a3fb7cc230e8ec3a2eefd2e896fe1be1a6a57668'}]}, 'timestamp': '2025-11-27 10:11:44.271003', '_unique_id': '46994d12ed63401f8c9a5abea656724e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:11:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:11:44.271 12 ERROR oslo_messaging.notify.messaging Nov 27 05:11:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:11:44 Nov 27 05:11:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:11:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:11:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['images', 'volumes', 'manila_metadata', 'vms', '.mgr', 'manila_data', 'backups'] Nov 27 05:11:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:11:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:11:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:11:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:11:44 localhost podman[328121]: 2025-11-27 10:11:44.71362281 +0000 UTC m=+0.081168472 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:11:44 localhost podman[328121]: 2025-11-27 10:11:44.72366751 +0000 UTC m=+0.091213212 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:11:44 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:11:44 localhost systemd[1]: tmp-crun.JXMs7m.mount: Deactivated successfully. Nov 27 05:11:44 localhost podman[328120]: 2025-11-27 10:11:44.787605138 +0000 UTC m=+0.156167617 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_managed=true, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) Nov 27 05:11:44 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:44.792 2 INFO neutron.agent.securitygroups_rpc [None req-54b492f4-6217-474c-9701-1be30bd8d6ed 05b9784fce164f068a57644f6d91d842 f0dbf7b1d0c74f4298b85cffda76f08a - - default default] Security group member updated ['d3602146-6d4f-48ba-8a93-3f9eecd02d01']#033[00m Nov 27 05:11:44 localhost podman[328120]: 2025-11-27 10:11:44.827560941 +0000 UTC m=+0.196123400 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, config_id=edpm, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, tcib_managed=true, org.label-schema.license=GPLv2) Nov 27 05:11:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e225 e225: 6 total, 6 up, 6 in Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014874720826353993 of space, bias 1.0, pg target 0.2969985924995347 quantized to 32 (current 32) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.635783082077052e-06 of space, bias 1.0, pg target 0.0003255208333333333 quantized to 32 (current 32) Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:11:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0001543088707426019 of space, bias 4.0, pg target 0.12282986111111112 quantized to 16 (current 16) Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:11:44 localhost podman[328122]: 2025-11-27 10:11:44.847810444 +0000 UTC m=+0.210413583 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, distribution-scope=public, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter, io.openshift.expose-services=, managed_by=edpm_ansible, name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, vcs-type=git, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., version=9.6, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64) Nov 27 05:11:44 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:11:44 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:44.858 265123 INFO neutron.agent.linux.ip_lib [None req-5849fa0e-421b-4985-bb5a-46a4d877b5a4 - - - - - -] Device tapa6f81314-28 cannot be used as it has no MAC address#033[00m Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:11:44 localhost podman[328122]: 2025-11-27 10:11:44.862935661 +0000 UTC m=+0.225538820 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vendor=Red Hat, Inc., io.buildah.version=1.33.7, managed_by=edpm_ansible, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., name=ubi9-minimal, vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6, maintainer=Red Hat, Inc., release=1755695350, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, container_name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, config_id=edpm, distribution-scope=public, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}) Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:11:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:11:44 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:11:44 localhost nova_compute[284026]: 2025-11-27 10:11:44.886 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:44 localhost kernel: device tapa6f81314-28 entered promiscuous mode Nov 27 05:11:44 localhost NetworkManager[5971]: [1764238304.8984] manager: (tapa6f81314-28): new Generic device (/org/freedesktop/NetworkManager/Devices/53) Nov 27 05:11:44 localhost systemd-udevd[328189]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:11:44 localhost nova_compute[284026]: 2025-11-27 10:11:44.902 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:44 localhost ovn_controller[156436]: 2025-11-27T10:11:44Z|00316|binding|INFO|Claiming lport a6f81314-288a-41de-9507-6d759019a879 for this chassis. Nov 27 05:11:44 localhost ovn_controller[156436]: 2025-11-27T10:11:44Z|00317|binding|INFO|a6f81314-288a-41de-9507-6d759019a879: Claiming unknown Nov 27 05:11:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:44.918 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f0dbf7b1d0c74f4298b85cffda76f08a', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=436bfce7-9f25-450f-bcc3-5a884c958354, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=a6f81314-288a-41de-9507-6d759019a879) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:44.919 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a6f81314-288a-41de-9507-6d759019a879 in datapath 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7 bound to our chassis#033[00m Nov 27 05:11:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:44.920 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:11:44 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:44.921 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[fc943483-c578-4e89-9099-a51fb5eec4b1]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:11:44 localhost ovn_controller[156436]: 2025-11-27T10:11:44Z|00318|binding|INFO|Setting lport a6f81314-288a-41de-9507-6d759019a879 ovn-installed in OVS Nov 27 05:11:44 localhost ovn_controller[156436]: 2025-11-27T10:11:44Z|00319|binding|INFO|Setting lport a6f81314-288a-41de-9507-6d759019a879 up in Southbound Nov 27 05:11:44 localhost nova_compute[284026]: 2025-11-27 10:11:44.944 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:44 localhost nova_compute[284026]: 2025-11-27 10:11:44.982 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:45 localhost nova_compute[284026]: 2025-11-27 10:11:45.016 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v500: 177 pgs: 177 active+clean; 194 MiB data, 984 MiB used, 41 GiB / 42 GiB avail; 38 KiB/s rd, 37 KiB/s wr, 60 op/s Nov 27 05:11:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume resize", "vol_name": "cephfs", "sub_name": "7260cd56-3d12-425d-bc6a-99686549f079", "new_size": 2147483648, "format": "json"}]: dispatch Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:11:45 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:11:45 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8/56c46d06-8c9b-4129-a913-44ae32f2140b Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:11:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:46 localhost podman[328245]: Nov 27 05:11:46 localhost podman[328245]: 2025-11-27 10:11:46.013372804 +0000 UTC m=+0.095894287 container create 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:11:46 localhost systemd[1]: Started libpod-conmon-93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80.scope. Nov 27 05:11:46 localhost podman[328245]: 2025-11-27 10:11:45.969172187 +0000 UTC m=+0.051693680 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:11:46 localhost nova_compute[284026]: 2025-11-27 10:11:46.114 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:46 localhost systemd[1]: Started libcrun container. Nov 27 05:11:46 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/0dd8bd16aad817d0da84b8d74be4b5db5a78378b703142a9ed0b55abdc091696/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:11:46 localhost podman[328245]: 2025-11-27 10:11:46.139291926 +0000 UTC m=+0.221813399 container init 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:11:46 localhost podman[328245]: 2025-11-27 10:11:46.150468426 +0000 UTC m=+0.232989899 container start 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:11:46 localhost dnsmasq[328264]: started, version 2.85 cachesize 150 Nov 27 05:11:46 localhost dnsmasq[328264]: DNS service limited to local subnets Nov 27 05:11:46 localhost dnsmasq[328264]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:11:46 localhost dnsmasq[328264]: warning: no upstream servers configured Nov 27 05:11:46 localhost dnsmasq-dhcp[328264]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:11:46 localhost dnsmasq[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/addn_hosts - 0 addresses Nov 27 05:11:46 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/host Nov 27 05:11:46 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/opts Nov 27 05:11:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "format": "json"}]: dispatch Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:46.201+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8' of type subvolume Nov 27 05:11:46 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8' of type subvolume Nov 27 05:11:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8", "force": true, "format": "json"}]: dispatch Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:46.215 265123 INFO neutron.agent.dhcp.agent [None req-5849fa0e-421b-4985-bb5a-46a4d877b5a4 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:11:44Z, description=, device_id=, device_owner=, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3fbc0d6c-4110-4453-8a08-f7542ef7a43a, ip_allocation=immediate, mac_address=fa:16:3e:6e:73:9c, name=tempest-RoutersIpV6Test-1101061770, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:11:42Z, description=, dns_domain=, id=9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-RoutersIpV6Test-2084563008, port_security_enabled=True, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=43975, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=3043, status=ACTIVE, subnets=['71f9f81e-b45b-4173-a2c7-91420a8cfc95'], tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:11:43Z, vlan_transparent=None, network_id=9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, port_security_enabled=True, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=['d3602146-6d4f-48ba-8a93-3f9eecd02d01'], standard_attr_id=3050, status=DOWN, tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:11:44Z on network 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7#033[00m Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8'' moved to trashcan Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ac96d349-e677-4ac8-bb9b-1a5a5b1d11e8, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta' Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "format": "json"}]: dispatch Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:46.355 265123 INFO neutron.agent.dhcp.agent [None req-b5815f8f-4bf9-4e9a-8627-57cbe3c66d70 - - - - - -] DHCP configuration for ports {'995f25b4-81d9-4432-8c42-7b2f48f9b8b6'} is completed#033[00m Nov 27 05:11:46 localhost dnsmasq[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/addn_hosts - 1 addresses Nov 27 05:11:46 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/host Nov 27 05:11:46 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/opts Nov 27 05:11:46 localhost podman[328283]: 2025-11-27 10:11:46.444954797 +0000 UTC m=+0.067558056 container kill 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:11:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:46.720 265123 INFO neutron.agent.dhcp.agent [None req-65d6335c-c3b5-4e7d-b448-91ab27e6f596 - - - - - -] DHCP configuration for ports {'3fbc0d6c-4110-4453-8a08-f7542ef7a43a'} is completed#033[00m Nov 27 05:11:46 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:46 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:46 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:46 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:11:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "96dccb75-d141-43d5-b863-c9f636d8f982", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:47 localhost nova_compute[284026]: 2025-11-27 10:11:47.323 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v501: 177 pgs: 177 active+clean; 194 MiB data, 1002 MiB used, 41 GiB / 42 GiB avail; 40 KiB/s rd, 35 KiB/s wr, 59 op/s Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/96dccb75-d141-43d5-b863-c9f636d8f982/.meta.tmp' Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/96dccb75-d141-43d5-b863-c9f636d8f982/.meta.tmp' to config b'/volumes/_nogroup/96dccb75-d141-43d5-b863-c9f636d8f982/.meta' Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "96dccb75-d141-43d5-b863-c9f636d8f982", "format": "json"}]: dispatch Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:47 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:47.362 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:11:44Z, description=, device_id=006b97b8-5445-41da-8b08-8b3d3d1fe8cb, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=3fbc0d6c-4110-4453-8a08-f7542ef7a43a, ip_allocation=immediate, mac_address=fa:16:3e:6e:73:9c, name=tempest-RoutersIpV6Test-1101061770, network_id=9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, port_security_enabled=True, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=3, security_groups=['d3602146-6d4f-48ba-8a93-3f9eecd02d01'], standard_attr_id=3050, status=ACTIVE, tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:11:46Z on network 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7#033[00m Nov 27 05:11:47 localhost dnsmasq[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/addn_hosts - 1 addresses Nov 27 05:11:47 localhost podman[328324]: 2025-11-27 10:11:47.567648875 +0000 UTC m=+0.061606586 container kill 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:11:47 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/host Nov 27 05:11:47 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/opts Nov 27 05:11:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e225 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:48 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:48.166 265123 INFO neutron.agent.dhcp.agent [None req-8ee908a5-46a8-48b2-aea4-e817449af79a - - - - - -] DHCP configuration for ports {'3fbc0d6c-4110-4453-8a08-f7542ef7a43a'} is completed#033[00m Nov 27 05:11:48 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:11:48 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:11:48 localhost podman[328360]: 2025-11-27 10:11:48.773186939 +0000 UTC m=+0.067734901 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:11:48 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:11:48 localhost systemd[1]: tmp-crun.9XFdg2.mount: Deactivated successfully. Nov 27 05:11:48 localhost ovn_controller[156436]: 2025-11-27T10:11:48Z|00320|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:11:48 localhost nova_compute[284026]: 2025-11-27 10:11:48.892 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "7260cd56-3d12-425d-bc6a-99686549f079", "format": "json"}]: dispatch Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:7260cd56-3d12-425d-bc6a-99686549f079, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:7260cd56-3d12-425d-bc6a-99686549f079, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7260cd56-3d12-425d-bc6a-99686549f079' of type subvolume Nov 27 05:11:49 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:49.186+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7260cd56-3d12-425d-bc6a-99686549f079' of type subvolume Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "7260cd56-3d12-425d-bc6a-99686549f079", "force": true, "format": "json"}]: dispatch Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/7260cd56-3d12-425d-bc6a-99686549f079'' moved to trashcan Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7260cd56-3d12-425d-bc6a-99686549f079, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v502: 177 pgs: 177 active+clean; 194 MiB data, 1002 MiB used, 41 GiB / 42 GiB avail; 37 KiB/s rd, 32 KiB/s wr, 54 op/s Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/.meta.tmp' Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/.meta.tmp' to config b'/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/.meta' Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "format": "json"}]: dispatch Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "snap_name": "9af2b852-b6c2-4333-abef-4284193a3f8d", "format": "json"}]: dispatch Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:49 localhost neutron_sriov_agent[258162]: 2025-11-27 10:11:49.806 2 INFO neutron.agent.securitygroups_rpc [None req-c3f629ef-a7df-4765-b1dd-a1c3db86f463 05b9784fce164f068a57644f6d91d842 f0dbf7b1d0c74f4298b85cffda76f08a - - default default] Security group member updated ['d3602146-6d4f-48ba-8a93-3f9eecd02d01']#033[00m Nov 27 05:11:50 localhost systemd[1]: tmp-crun.UVddE9.mount: Deactivated successfully. Nov 27 05:11:50 localhost podman[328397]: 2025-11-27 10:11:50.035750365 +0000 UTC m=+0.071719777 container kill 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125) Nov 27 05:11:50 localhost dnsmasq[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/addn_hosts - 0 addresses Nov 27 05:11:50 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/host Nov 27 05:11:50 localhost dnsmasq-dhcp[328264]: read /var/lib/neutron/dhcp/9a48e224-c8ad-420a-ba4c-73bef2aa8aa7/opts Nov 27 05:11:50 localhost nova_compute[284026]: 2025-11-27 10:11:50.273 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:50 localhost kernel: device tapa6f81314-28 left promiscuous mode Nov 27 05:11:50 localhost ovn_controller[156436]: 2025-11-27T10:11:50Z|00321|binding|INFO|Releasing lport a6f81314-288a-41de-9507-6d759019a879 from this chassis (sb_readonly=0) Nov 27 05:11:50 localhost ovn_controller[156436]: 2025-11-27T10:11:50Z|00322|binding|INFO|Setting lport a6f81314-288a-41de-9507-6d759019a879 down in Southbound Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.296 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f0dbf7b1d0c74f4298b85cffda76f08a', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=436bfce7-9f25-450f-bcc3-5a884c958354, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=a6f81314-288a-41de-9507-6d759019a879) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.298 162092 INFO neutron.agent.ovn.metadata.agent [-] Port a6f81314-288a-41de-9507-6d759019a879 in datapath 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7 unbound from our chassis#033[00m Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.300 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 9a48e224-c8ad-420a-ba4c-73bef2aa8aa7 or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.301 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[3620dc91-9a8a-4fb5-a810-14b0dc3adab6]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:11:50 localhost nova_compute[284026]: 2025-11-27 10:11:50.303 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.509 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=22, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=21) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:11:50 localhost nova_compute[284026]: 2025-11-27 10:11:50.510 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:50 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:50.511 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 1 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:11:51 localhost nova_compute[284026]: 2025-11-27 10:11:51.117 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v503: 177 pgs: 177 active+clean; 194 MiB data, 1003 MiB used, 41 GiB / 42 GiB avail; 31 KiB/s rd, 69 KiB/s wr, 52 op/s Nov 27 05:11:51 localhost ovn_metadata_agent[162087]: 2025-11-27 10:11:51.514 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '22'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:11:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "96dccb75-d141-43d5-b863-c9f636d8f982", "format": "json"}]: dispatch Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:96dccb75-d141-43d5-b863-c9f636d8f982, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:96dccb75-d141-43d5-b863-c9f636d8f982, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:51 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '96dccb75-d141-43d5-b863-c9f636d8f982' of type subvolume Nov 27 05:11:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:51.560+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '96dccb75-d141-43d5-b863-c9f636d8f982' of type subvolume Nov 27 05:11:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "96dccb75-d141-43d5-b863-c9f636d8f982", "force": true, "format": "json"}]: dispatch Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/96dccb75-d141-43d5-b863-c9f636d8f982'' moved to trashcan Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:51 localhost dnsmasq[328264]: exiting on receipt of SIGTERM Nov 27 05:11:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:96dccb75-d141-43d5-b863-c9f636d8f982, vol_name:cephfs) < "" Nov 27 05:11:51 localhost podman[328434]: 2025-11-27 10:11:51.586688296 +0000 UTC m=+0.073950398 container kill 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:11:51 localhost systemd[1]: libpod-93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80.scope: Deactivated successfully. Nov 27 05:11:51 localhost podman[328450]: 2025-11-27 10:11:51.664168957 +0000 UTC m=+0.049393997 container died 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:11:51 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80-userdata-shm.mount: Deactivated successfully. Nov 27 05:11:51 localhost systemd[1]: var-lib-containers-storage-overlay-0dd8bd16aad817d0da84b8d74be4b5db5a78378b703142a9ed0b55abdc091696-merged.mount: Deactivated successfully. Nov 27 05:11:51 localhost podman[328450]: 2025-11-27 10:11:51.712995729 +0000 UTC m=+0.098220759 container remove 93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-9a48e224-c8ad-420a-ba4c-73bef2aa8aa7, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:11:51 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:51.756 265123 INFO neutron.agent.dhcp.agent [None req-db9d9847-abfc-40c4-b84d-58830f74c1ee - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:11:51 localhost systemd[1]: run-netns-qdhcp\x2d9a48e224\x2dc8ad\x2d420a\x2dba4c\x2d73bef2aa8aa7.mount: Deactivated successfully. Nov 27 05:11:51 localhost systemd[1]: libpod-conmon-93128b23b1be4808cfc8bf66bb1f5850fc6005da715fd9f811cd9750a5ed1b80.scope: Deactivated successfully. Nov 27 05:11:51 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:11:51.872 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:11:52 localhost ovn_controller[156436]: 2025-11-27T10:11:52Z|00323|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:11:52 localhost nova_compute[284026]: 2025-11-27 10:11:52.091 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "09a50d5b-2cd2-4b70-a273-25b86623ef56", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/09a50d5b-2cd2-4b70-a273-25b86623ef56/.meta.tmp' Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/09a50d5b-2cd2-4b70-a273-25b86623ef56/.meta.tmp' to config b'/volumes/_nogroup/09a50d5b-2cd2-4b70-a273-25b86623ef56/.meta' Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "09a50d5b-2cd2-4b70-a273-25b86623ef56", "format": "json"}]: dispatch Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:52 localhost nova_compute[284026]: 2025-11-27 10:11:52.325 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e226 e226: 6 total, 6 up, 6 in Nov 27 05:11:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:11:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:11:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:11:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b", "osd", "allow rw pool=manila_data namespace=fsvolumens_aa05bb5f-0d8b-4154-9620-09e41c5b956b", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:11:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b", "osd", "allow rw pool=manila_data namespace=fsvolumens_aa05bb5f-0d8b-4154-9620-09e41c5b956b", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:11:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "snap_name": "9af2b852-b6c2-4333-abef-4284193a3f8d_4c60878d-f585-40d8-81a6-94f073156ef0", "force": true, "format": "json"}]: dispatch Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d_4c60878d-f585-40d8-81a6-94f073156ef0, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta' Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d_4c60878d-f585-40d8-81a6-94f073156ef0, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "snap_name": "9af2b852-b6c2-4333-abef-4284193a3f8d", "force": true, "format": "json"}]: dispatch Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta.tmp' to config b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf/.meta' Nov 27 05:11:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e226 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:9af2b852-b6c2-4333-abef-4284193a3f8d, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v505: 177 pgs: 177 active+clean; 194 MiB data, 1003 MiB used, 41 GiB / 42 GiB avail; 29 KiB/s rd, 65 KiB/s wr, 49 op/s Nov 27 05:11:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b", "osd", "allow rw pool=manila_data namespace=fsvolumens_aa05bb5f-0d8b-4154-9620-09e41c5b956b", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b", "osd", "allow rw pool=manila_data namespace=fsvolumens_aa05bb5f-0d8b-4154-9620-09e41c5b956b", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:11:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b", "osd", "allow rw pool=manila_data namespace=fsvolumens_aa05bb5f-0d8b-4154-9620-09e41c5b956b", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:11:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:11:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:11:53 localhost podman[328476]: 2025-11-27 10:11:53.990208759 +0000 UTC m=+0.086940046 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=ovn_controller, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller) Nov 27 05:11:54 localhost podman[328476]: 2025-11-27 10:11:54.025985271 +0000 UTC m=+0.122716618 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:11:54 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:11:54 localhost podman[328477]: 2025-11-27 10:11:54.064558616 +0000 UTC m=+0.158852727 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:11:54 localhost podman[328477]: 2025-11-27 10:11:54.079059375 +0000 UTC m=+0.173353516 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:11:54 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:11:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "e81b59ff-73ed-4717-af17-96e0fedb08de", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/e81b59ff-73ed-4717-af17-96e0fedb08de/.meta.tmp' Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/e81b59ff-73ed-4717-af17-96e0fedb08de/.meta.tmp' to config b'/volumes/_nogroup/e81b59ff-73ed-4717-af17-96e0fedb08de/.meta' Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "e81b59ff-73ed-4717-af17-96e0fedb08de", "format": "json"}]: dispatch Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v506: 177 pgs: 177 active+clean; 194 MiB data, 1003 MiB used, 41 GiB / 42 GiB avail; 24 KiB/s rd, 55 KiB/s wr, 42 op/s Nov 27 05:11:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "09a50d5b-2cd2-4b70-a273-25b86623ef56", "format": "json"}]: dispatch Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '09a50d5b-2cd2-4b70-a273-25b86623ef56' of type subvolume Nov 27 05:11:55 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:55.465+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '09a50d5b-2cd2-4b70-a273-25b86623ef56' of type subvolume Nov 27 05:11:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "09a50d5b-2cd2-4b70-a273-25b86623ef56", "force": true, "format": "json"}]: dispatch Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/09a50d5b-2cd2-4b70-a273-25b86623ef56'' moved to trashcan Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:09a50d5b-2cd2-4b70-a273-25b86623ef56, vol_name:cephfs) < "" Nov 27 05:11:55 localhost openstack_network_exporter[244641]: ERROR 10:11:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:11:55 localhost openstack_network_exporter[244641]: ERROR 10:11:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:11:55 localhost openstack_network_exporter[244641]: ERROR 10:11:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:11:55 localhost openstack_network_exporter[244641]: ERROR 10:11:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:11:55 localhost openstack_network_exporter[244641]: Nov 27 05:11:55 localhost openstack_network_exporter[244641]: ERROR 10:11:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:11:55 localhost openstack_network_exporter[244641]: Nov 27 05:11:56 localhost nova_compute[284026]: 2025-11-27 10:11:56.120 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '2c0fce53-32ef-4091-a6e4-901cdb7e77cf' of type subvolume Nov 27 05:11:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:56.135+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '2c0fce53-32ef-4091-a6e4-901cdb7e77cf' of type subvolume Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "2c0fce53-32ef-4091-a6e4-901cdb7e77cf", "force": true, "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/2c0fce53-32ef-4091-a6e4-901cdb7e77cf'' moved to trashcan Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:2c0fce53-32ef-4091-a6e4-901cdb7e77cf, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:11:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:11:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b/94883c59-e03d-4a68-af00-e950bf93896b Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'aa05bb5f-0d8b-4154-9620-09e41c5b956b' of type subvolume Nov 27 05:11:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:56.602+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'aa05bb5f-0d8b-4154-9620-09e41c5b956b' of type subvolume Nov 27 05:11:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "aa05bb5f-0d8b-4154-9620-09e41c5b956b", "force": true, "format": "json"}]: dispatch Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/aa05bb5f-0d8b-4154-9620-09e41c5b956b'' moved to trashcan Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:aa05bb5f-0d8b-4154-9620-09e41c5b956b, vol_name:cephfs) < "" Nov 27 05:11:56 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:11:56 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:56 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:11:56 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:11:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:11:56 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1780660274' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:11:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:11:56 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1780660274' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:11:57 localhost nova_compute[284026]: 2025-11-27 10:11:57.329 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:11:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v507: 177 pgs: 177 active+clean; 195 MiB data, 1004 MiB used, 41 GiB / 42 GiB avail; 1.3 KiB/s rd, 68 KiB/s wr, 13 op/s Nov 27 05:11:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e227 e227: 6 total, 6 up, 6 in Nov 27 05:11:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e227 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "e81b59ff-73ed-4717-af17-96e0fedb08de", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:e81b59ff-73ed-4717-af17-96e0fedb08de, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:e81b59ff-73ed-4717-af17-96e0fedb08de, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:11:59.084+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e81b59ff-73ed-4717-af17-96e0fedb08de' of type subvolume Nov 27 05:11:59 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e81b59ff-73ed-4717-af17-96e0fedb08de' of type subvolume Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "e81b59ff-73ed-4717-af17-96e0fedb08de", "force": true, "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/e81b59ff-73ed-4717-af17-96e0fedb08de'' moved to trashcan Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e81b59ff-73ed-4717-af17-96e0fedb08de, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "557eb994-7270-4607-9e22-ed22bad9ab54", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v509: 177 pgs: 177 active+clean; 195 MiB data, 1004 MiB used, 41 GiB / 42 GiB avail; 1.2 KiB/s rd, 43 KiB/s wr, 8 op/s Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/557eb994-7270-4607-9e22-ed22bad9ab54/.meta.tmp' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/557eb994-7270-4607-9e22-ed22bad9ab54/.meta.tmp' to config b'/volumes/_nogroup/557eb994-7270-4607-9e22-ed22bad9ab54/.meta' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "557eb994-7270-4607-9e22-ed22bad9ab54", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/.meta.tmp' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/.meta.tmp' to config b'/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/.meta' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "5261ca91-accf-4ae6-89f7-1ab840580f10", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/5261ca91-accf-4ae6-89f7-1ab840580f10/.meta.tmp' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/5261ca91-accf-4ae6-89f7-1ab840580f10/.meta.tmp' to config b'/volumes/_nogroup/5261ca91-accf-4ae6-89f7-1ab840580f10/.meta' Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "5261ca91-accf-4ae6-89f7-1ab840580f10", "format": "json"}]: dispatch Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:11:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:12:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:00.246 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:11:59Z, description=, device_id=1dd366a6-c750-4ab1-aae3-2d207eb6dc51, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e98344bb-b552-4799-9e35-ebea33bdcc20, ip_allocation=immediate, mac_address=fa:16:3e:9c:fe:86, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3072, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:12:00Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:12:00 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:12:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:12:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:12:00 localhost podman[328542]: 2025-11-27 10:12:00.467615026 +0000 UTC m=+0.054432103 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:12:00 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:00.963 265123 INFO neutron.agent.dhcp.agent [None req-2d65542a-3f16-45c8-ad6d-d2d70774efbd - - - - - -] DHCP configuration for ports {'e98344bb-b552-4799-9e35-ebea33bdcc20'} is completed#033[00m Nov 27 05:12:01 localhost nova_compute[284026]: 2025-11-27 10:12:01.122 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v510: 177 pgs: 177 active+clean; 195 MiB data, 1009 MiB used, 41 GiB / 42 GiB avail; 16 KiB/s rd, 92 KiB/s wr, 35 op/s Nov 27 05:12:02 localhost nova_compute[284026]: 2025-11-27 10:12:02.333 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "986ef845-7a7d-4202-945b-6651c46fc4d8", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/986ef845-7a7d-4202-945b-6651c46fc4d8/.meta.tmp' Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/986ef845-7a7d-4202-945b-6651c46fc4d8/.meta.tmp' to config b'/volumes/_nogroup/986ef845-7a7d-4202-945b-6651c46fc4d8/.meta' Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "986ef845-7a7d-4202-945b-6651c46fc4d8", "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e228 e228: 6 total, 6 up, 6 in Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "557eb994-7270-4607-9e22-ed22bad9ab54", "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:557eb994-7270-4607-9e22-ed22bad9ab54, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:557eb994-7270-4607-9e22-ed22bad9ab54, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:02.765+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '557eb994-7270-4607-9e22-ed22bad9ab54' of type subvolume Nov 27 05:12:02 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '557eb994-7270-4607-9e22-ed22bad9ab54' of type subvolume Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "557eb994-7270-4607-9e22-ed22bad9ab54", "force": true, "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/557eb994-7270-4607-9e22-ed22bad9ab54'' moved to trashcan Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:557eb994-7270-4607-9e22-ed22bad9ab54, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "5261ca91-accf-4ae6-89f7-1ab840580f10", "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:5261ca91-accf-4ae6-89f7-1ab840580f10, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:5261ca91-accf-4ae6-89f7-1ab840580f10, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:02.957+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5261ca91-accf-4ae6-89f7-1ab840580f10' of type subvolume Nov 27 05:12:02 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5261ca91-accf-4ae6-89f7-1ab840580f10' of type subvolume Nov 27 05:12:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "5261ca91-accf-4ae6-89f7-1ab840580f10", "force": true, "format": "json"}]: dispatch Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/5261ca91-accf-4ae6-89f7-1ab840580f10'' moved to trashcan Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5261ca91-accf-4ae6-89f7-1ab840580f10, vol_name:cephfs) < "" Nov 27 05:12:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:12:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:ef540655-822a-4748-b398-a773bda561db, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:03 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:12:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e228 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537", "osd", "allow rw pool=manila_data namespace=fsvolumens_ef540655-822a-4748-b398-a773bda561db", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:12:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537", "osd", "allow rw pool=manila_data namespace=fsvolumens_ef540655-822a-4748-b398-a773bda561db", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:ef540655-822a-4748-b398-a773bda561db, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v512: 177 pgs: 177 active+clean; 195 MiB data, 1009 MiB used, 41 GiB / 42 GiB avail; 17 KiB/s rd, 99 KiB/s wr, 37 op/s Nov 27 05:12:03 localhost neutron_sriov_agent[258162]: 2025-11-27 10:12:03.821 2 INFO neutron.agent.securitygroups_rpc [None req-441c1baf-e0ea-4412-bc95-e4697b934b92 db93d12e5a1e42be891b77305017763c 50343d8ff77a4435ae1c80cd4e9fe260 - - default default] Security group member updated ['eb8b7b07-030c-4ea1-abbc-8c59c1d757eb']#033[00m Nov 27 05:12:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537", "osd", "allow rw pool=manila_data namespace=fsvolumens_ef540655-822a-4748-b398-a773bda561db", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537", "osd", "allow rw pool=manila_data namespace=fsvolumens_ef540655-822a-4748-b398-a773bda561db", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537", "osd", "allow rw pool=manila_data namespace=fsvolumens_ef540655-822a-4748-b398-a773bda561db", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:12:04 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:12:04 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:12:04 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:12:04 localhost podman[328579]: 2025-11-27 10:12:04.072378288 +0000 UTC m=+0.066734384 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:12:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v513: 177 pgs: 177 active+clean; 195 MiB data, 1009 MiB used, 41 GiB / 42 GiB avail; 16 KiB/s rd, 56 KiB/s wr, 28 op/s Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "afd5c026-3df5-4c59-bab1-3fed41bb9f9d", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:06 localhost nova_compute[284026]: 2025-11-27 10:12:06.125 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta' Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:12:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db/9abf868a-c260-4aee-83c5-fe1961a22537 Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:06 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:06 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:12:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:12:06 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:12:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "2c9f01d8-a144-4229-8359-c5f872971db6", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:06 localhost systemd[1]: tmp-crun.iMxLut.mount: Deactivated successfully. Nov 27 05:12:06 localhost podman[328603]: 2025-11-27 10:12:06.99479273 +0000 UTC m=+0.085993471 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_id=multipathd, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, tcib_managed=true) Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/2c9f01d8-a144-4229-8359-c5f872971db6/.meta.tmp' Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/2c9f01d8-a144-4229-8359-c5f872971db6/.meta.tmp' to config b'/volumes/_nogroup/2c9f01d8-a144-4229-8359-c5f872971db6/.meta' Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:07 localhost podman[328603]: 2025-11-27 10:12:07.035984747 +0000 UTC m=+0.127185498 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, container_name=multipathd, io.buildah.version=1.41.3, tcib_managed=true, org.label-schema.vendor=CentOS, managed_by=edpm_ansible) Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "2c9f01d8-a144-4229-8359-c5f872971db6", "format": "json"}]: dispatch Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:07 localhost podman[328602]: 2025-11-27 10:12:07.050406884 +0000 UTC m=+0.142841178 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent) Nov 27 05:12:07 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:12:07 localhost podman[328602]: 2025-11-27 10:12:07.059592691 +0000 UTC m=+0.152026955 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible) Nov 27 05:12:07 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "ef540655-822a-4748-b398-a773bda561db", "format": "json"}]: dispatch Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:ef540655-822a-4748-b398-a773bda561db, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:ef540655-822a-4748-b398-a773bda561db, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:07.153+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ef540655-822a-4748-b398-a773bda561db' of type subvolume Nov 27 05:12:07 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ef540655-822a-4748-b398-a773bda561db' of type subvolume Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "ef540655-822a-4748-b398-a773bda561db", "force": true, "format": "json"}]: dispatch Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/ef540655-822a-4748-b398-a773bda561db'' moved to trashcan Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ef540655-822a-4748-b398-a773bda561db, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "986ef845-7a7d-4202-945b-6651c46fc4d8", "format": "json"}]: dispatch Nov 27 05:12:07 localhost neutron_sriov_agent[258162]: 2025-11-27 10:12:07.270 2 INFO neutron.agent.securitygroups_rpc [None req-6d30c411-d2fe-41db-93fb-af8e78e5f9f2 db93d12e5a1e42be891b77305017763c 50343d8ff77a4435ae1c80cd4e9fe260 - - default default] Security group member updated ['eb8b7b07-030c-4ea1-abbc-8c59c1d757eb']#033[00m Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:986ef845-7a7d-4202-945b-6651c46fc4d8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:986ef845-7a7d-4202-945b-6651c46fc4d8, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:07.277+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '986ef845-7a7d-4202-945b-6651c46fc4d8' of type subvolume Nov 27 05:12:07 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '986ef845-7a7d-4202-945b-6651c46fc4d8' of type subvolume Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "986ef845-7a7d-4202-945b-6651c46fc4d8", "force": true, "format": "json"}]: dispatch Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/986ef845-7a7d-4202-945b-6651c46fc4d8'' moved to trashcan Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:07 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:986ef845-7a7d-4202-945b-6651c46fc4d8, vol_name:cephfs) < "" Nov 27 05:12:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v514: 177 pgs: 177 active+clean; 195 MiB data, 1010 MiB used, 41 GiB / 42 GiB avail; 14 KiB/s rd, 73 KiB/s wr, 28 op/s Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.357 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:07 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:07.397 265123 INFO neutron.agent.linux.ip_lib [None req-c7a1a4f0-86a4-4287-82e4-b097c162b295 - - - - - -] Device tap4141b727-a3 cannot be used as it has no MAC address#033[00m Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.423 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:07 localhost kernel: device tap4141b727-a3 entered promiscuous mode Nov 27 05:12:07 localhost NetworkManager[5971]: [1764238327.4331] manager: (tap4141b727-a3): new Generic device (/org/freedesktop/NetworkManager/Devices/54) Nov 27 05:12:07 localhost ovn_controller[156436]: 2025-11-27T10:12:07Z|00324|binding|INFO|Claiming lport 4141b727-a354-4bc7-968b-e9cd93fc1c32 for this chassis. Nov 27 05:12:07 localhost ovn_controller[156436]: 2025-11-27T10:12:07Z|00325|binding|INFO|4141b727-a354-4bc7-968b-e9cd93fc1c32: Claiming unknown Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.433 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:07 localhost systemd-udevd[328650]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:12:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:07.446 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-38135c0c-47a1-4390-8c4d-35ca5410096a', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-38135c0c-47a1-4390-8c4d-35ca5410096a', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f0dbf7b1d0c74f4298b85cffda76f08a', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=19da2a90-4bfc-4717-bdc8-2d0144f41b35, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4141b727-a354-4bc7-968b-e9cd93fc1c32) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:12:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:07.448 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4141b727-a354-4bc7-968b-e9cd93fc1c32 in datapath 38135c0c-47a1-4390-8c4d-35ca5410096a bound to our chassis#033[00m Nov 27 05:12:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:07.450 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 38135c0c-47a1-4390-8c4d-35ca5410096a or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:12:07 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:07.451 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[49a86aab-fcbf-47fd-ac3e-0e8dfe0e6a48]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost ovn_controller[156436]: 2025-11-27T10:12:07Z|00326|binding|INFO|Setting lport 4141b727-a354-4bc7-968b-e9cd93fc1c32 ovn-installed in OVS Nov 27 05:12:07 localhost ovn_controller[156436]: 2025-11-27T10:12:07Z|00327|binding|INFO|Setting lport 4141b727-a354-4bc7-968b-e9cd93fc1c32 up in Southbound Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.477 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost journal[232028]: ethtool ioctl error on tap4141b727-a3: No such device Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.517 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:07 localhost nova_compute[284026]: 2025-11-27 10:12:07.547 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e228 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:08 localhost podman[328757]: Nov 27 05:12:08 localhost podman[328757]: 2025-11-27 10:12:08.395027194 +0000 UTC m=+0.093773481 container create c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:12:08 localhost systemd[1]: Started libpod-conmon-c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215.scope. Nov 27 05:12:08 localhost podman[328757]: 2025-11-27 10:12:08.351107254 +0000 UTC m=+0.049853541 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:12:08 localhost systemd[1]: Started libcrun container. Nov 27 05:12:08 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/911c17bfaa27c89bad88a1b506a75e32be279f16ba1fd41802de2e8dd0528637/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:12:08 localhost podman[328757]: 2025-11-27 10:12:08.47343092 +0000 UTC m=+0.172177217 container init c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:12:08 localhost podman[328757]: 2025-11-27 10:12:08.484663362 +0000 UTC m=+0.183409649 container start c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) Nov 27 05:12:08 localhost dnsmasq[328776]: started, version 2.85 cachesize 150 Nov 27 05:12:08 localhost dnsmasq[328776]: DNS service limited to local subnets Nov 27 05:12:08 localhost dnsmasq[328776]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:12:08 localhost dnsmasq[328776]: warning: no upstream servers configured Nov 27 05:12:08 localhost dnsmasq-dhcp[328776]: DHCPv6, static leases only on 2001:db8::, lease time 1d Nov 27 05:12:08 localhost dnsmasq[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/addn_hosts - 0 addresses Nov 27 05:12:08 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/host Nov 27 05:12:08 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/opts Nov 27 05:12:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:08.548 265123 INFO neutron.agent.dhcp.agent [None req-c7a1a4f0-86a4-4287-82e4-b097c162b295 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:12:06Z, description=, device_id=dd04de78-21d2-4390-8d59-a6a7ecd45712, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e6bd5bad-1a3a-4349-a79b-b803df4473a4, ip_allocation=immediate, mac_address=fa:16:3e:97:71:54, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:12:04Z, description=, dns_domain=, id=38135c0c-47a1-4390-8c4d-35ca5410096a, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-RoutersIpV6Test-1810273209, port_security_enabled=True, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=27165, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=3077, status=ACTIVE, subnets=['732c6b86-23d2-4484-b0b6-54ec7c5e1bdb'], tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:12:06Z, vlan_transparent=None, network_id=38135c0c-47a1-4390-8c4d-35ca5410096a, port_security_enabled=False, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3083, status=DOWN, tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:12:07Z on network 38135c0c-47a1-4390-8c4d-35ca5410096a#033[00m Nov 27 05:12:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:08.656 265123 INFO neutron.agent.dhcp.agent [None req-bfab3974-6f72-469c-9d56-6686ac5e319c - - - - - -] DHCP configuration for ports {'8c463a3b-0119-4e40-9a8a-f28142d4b5a0'} is completed#033[00m Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 05:12:08 localhost podman[242678]: time="2025-11-27T10:12:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 05:12:08 localhost dnsmasq[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/addn_hosts - 1 addresses Nov 27 05:12:08 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/host Nov 27 05:12:08 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/opts Nov 27 05:12:08 localhost podman[328814]: 2025-11-27 10:12:08.775143124 +0000 UTC m=+0.077906593 container kill c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:12:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 05:12:08 localhost podman[242678]: @ - - [27/Nov/2025:10:12:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 159329 "" "Go-http-client/1.1" Nov 27 05:12:08 localhost podman[242678]: @ - - [27/Nov/2025:10:12:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 20236 "" "Go-http-client/1.1" Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:08 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:08.962 265123 INFO neutron.agent.dhcp.agent [None req-c7a1a4f0-86a4-4287-82e4-b097c162b295 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:12:06Z, description=, device_id=dd04de78-21d2-4390-8d59-a6a7ecd45712, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e6bd5bad-1a3a-4349-a79b-b803df4473a4, ip_allocation=immediate, mac_address=fa:16:3e:97:71:54, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:12:04Z, description=, dns_domain=, id=38135c0c-47a1-4390-8c4d-35ca5410096a, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-RoutersIpV6Test-1810273209, port_security_enabled=True, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=27165, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=3077, status=ACTIVE, subnets=['732c6b86-23d2-4484-b0b6-54ec7c5e1bdb'], tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:12:06Z, vlan_transparent=None, network_id=38135c0c-47a1-4390-8c4d-35ca5410096a, port_security_enabled=False, project_id=f0dbf7b1d0c74f4298b85cffda76f08a, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3083, status=DOWN, tags=[], tenant_id=f0dbf7b1d0c74f4298b85cffda76f08a, updated_at=2025-11-27T10:12:07Z on network 38135c0c-47a1-4390-8c4d-35ca5410096a#033[00m Nov 27 05:12:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:09.115 265123 INFO neutron.agent.dhcp.agent [None req-7a33cfc7-b0e0-46f3-80d9-58b974b2bcce - - - - - -] DHCP configuration for ports {'e6bd5bad-1a3a-4349-a79b-b803df4473a4'} is completed#033[00m Nov 27 05:12:09 localhost podman[328889]: 2025-11-27 10:12:09.165837249 +0000 UTC m=+0.066648951 container kill c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:12:09 localhost dnsmasq[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/addn_hosts - 1 addresses Nov 27 05:12:09 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/host Nov 27 05:12:09 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/opts Nov 27 05:12:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v515: 177 pgs: 177 active+clean; 195 MiB data, 1010 MiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 69 KiB/s wr, 27 op/s Nov 27 05:12:09 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:09.462 265123 INFO neutron.agent.dhcp.agent [None req-924055f6-0df9-47b3-a185-834a7afeb817 - - - - - -] DHCP configuration for ports {'e6bd5bad-1a3a-4349-a79b-b803df4473a4'} is completed#033[00m Nov 27 05:12:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:12:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #44. Immutable memtables: 1. Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:12:09 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 746011ac-1b80-4399-b8aa-238a1442b5d0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:12:09 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 746011ac-1b80-4399-b8aa-238a1442b5d0 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:12:09 localhost ceph-mgr[290377]: [progress INFO root] Completed event 746011ac-1b80-4399-b8aa-238a1442b5d0 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:09 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:12:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:09 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:12:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "afd5c026-3df5-4c59-bab1-3fed41bb9f9d_0d782151-db2c-4bb7-b968-b23025738ff3", "force": true, "format": "json"}]: dispatch Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d_0d782151-db2c-4bb7-b968-b23025738ff3, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d_0d782151-db2c-4bb7-b968-b23025738ff3, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "afd5c026-3df5-4c59-bab1-3fed41bb9f9d", "force": true, "format": "json"}]: dispatch Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:afd5c026-3df5-4c59-bab1-3fed41bb9f9d, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "snap_name": "cf40cff8-dc34-46a6-ac29-9b5a499444e5", "format": "json"}]: dispatch Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:10 localhost podman[328976]: 2025-11-27 10:12:10.928865028 +0000 UTC m=+0.058839941 container kill c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0) Nov 27 05:12:10 localhost dnsmasq[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/addn_hosts - 0 addresses Nov 27 05:12:10 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/host Nov 27 05:12:10 localhost dnsmasq-dhcp[328776]: read /var/lib/neutron/dhcp/38135c0c-47a1-4390-8c4d-35ca5410096a/opts Nov 27 05:12:11 localhost nova_compute[284026]: 2025-11-27 10:12:11.138 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:11 localhost ovn_controller[156436]: 2025-11-27T10:12:11Z|00328|binding|INFO|Releasing lport 4141b727-a354-4bc7-968b-e9cd93fc1c32 from this chassis (sb_readonly=0) Nov 27 05:12:11 localhost ovn_controller[156436]: 2025-11-27T10:12:11Z|00329|binding|INFO|Setting lport 4141b727-a354-4bc7-968b-e9cd93fc1c32 down in Southbound Nov 27 05:12:11 localhost kernel: device tap4141b727-a3 left promiscuous mode Nov 27 05:12:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:11.158 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-38135c0c-47a1-4390-8c4d-35ca5410096a', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-38135c0c-47a1-4390-8c4d-35ca5410096a', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'f0dbf7b1d0c74f4298b85cffda76f08a', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=19da2a90-4bfc-4717-bdc8-2d0144f41b35, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=4141b727-a354-4bc7-968b-e9cd93fc1c32) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:12:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:11.160 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 4141b727-a354-4bc7-968b-e9cd93fc1c32 in datapath 38135c0c-47a1-4390-8c4d-35ca5410096a unbound from our chassis#033[00m Nov 27 05:12:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:11.161 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network 38135c0c-47a1-4390-8c4d-35ca5410096a or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:12:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:11.162 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[4e45f160-b44b-410d-a594-7a4d9c5ff0e3]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:12:11 localhost nova_compute[284026]: 2025-11-27 10:12:11.164 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v516: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 818 B/s rd, 77 KiB/s wr, 12 op/s Nov 27 05:12:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "2c9f01d8-a144-4229-8359-c5f872971db6", "format": "json"}]: dispatch Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:2c9f01d8-a144-4229-8359-c5f872971db6, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:12 localhost nova_compute[284026]: 2025-11-27 10:12:12.388 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:2c9f01d8-a144-4229-8359-c5f872971db6, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:12 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:12.391+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '2c9f01d8-a144-4229-8359-c5f872971db6' of type subvolume Nov 27 05:12:12 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '2c9f01d8-a144-4229-8359-c5f872971db6' of type subvolume Nov 27 05:12:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "2c9f01d8-a144-4229-8359-c5f872971db6", "force": true, "format": "json"}]: dispatch Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/2c9f01d8-a144-4229-8359-c5f872971db6'' moved to trashcan Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:2c9f01d8-a144-4229-8359-c5f872971db6, vol_name:cephfs) < "" Nov 27 05:12:12 localhost dnsmasq[328776]: exiting on receipt of SIGTERM Nov 27 05:12:12 localhost podman[329016]: 2025-11-27 10:12:12.746046102 +0000 UTC m=+0.059277053 container kill c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:12:12 localhost systemd[1]: libpod-c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215.scope: Deactivated successfully. Nov 27 05:12:12 localhost podman[329029]: 2025-11-27 10:12:12.834136388 +0000 UTC m=+0.068683255 container died c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:12:12 localhost systemd[1]: tmp-crun.rBsiO0.mount: Deactivated successfully. Nov 27 05:12:12 localhost podman[329029]: 2025-11-27 10:12:12.880075473 +0000 UTC m=+0.114622290 container cleanup c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:12:12 localhost systemd[1]: libpod-conmon-c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215.scope: Deactivated successfully. Nov 27 05:12:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "e7a60627-9c7e-4350-baca-e4831df56001", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/e7a60627-9c7e-4350-baca-e4831df56001/.meta.tmp' Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/e7a60627-9c7e-4350-baca-e4831df56001/.meta.tmp' to config b'/volumes/_nogroup/e7a60627-9c7e-4350-baca-e4831df56001/.meta' Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:12 localhost podman[329031]: 2025-11-27 10:12:12.962139167 +0000 UTC m=+0.190993961 container remove c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38135c0c-47a1-4390-8c4d-35ca5410096a, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.vendor=CentOS) Nov 27 05:12:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "e7a60627-9c7e-4350-baca-e4831df56001", "format": "json"}]: dispatch Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e229 e229: 6 total, 6 up, 6 in Nov 27 05:12:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e229 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:13.288 265123 INFO neutron.agent.dhcp.agent [None req-288afd81-f573-4c13-801c-c04c7c63da06 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:12:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:13.289 265123 INFO neutron.agent.dhcp.agent [None req-288afd81-f573-4c13-801c-c04c7c63da06 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:12:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v518: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 818 B/s rd, 77 KiB/s wr, 12 op/s Nov 27 05:12:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot clone", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "snap_name": "cf40cff8-dc34-46a6-ac29-9b5a499444e5", "target_sub_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "format": "json"}]: dispatch Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, target_sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 273 bytes to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] tracking-id 00cd5267-9c84-40d7-9be4-8a116f4f3d38 for path b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 246 bytes to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, target_sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.640+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.640+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.640+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.640+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.640+0000 7f54e6962640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "format": "json"}]: dispatch Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] cloning to subvolume path: /volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31 Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] starting clone: (cephfs, None, 24b71b24-5036-4167-9c06-e5bfad964f31) Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.667+0000 7f54e5960640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.667+0000 7f54e5960640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.667+0000 7f54e5960640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.667+0000 7f54e5960640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:13.667+0000 7f54e5960640 -1 client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: client.0 error registering admin socket command: (17) File exists Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] Delayed cloning (cephfs, None, 24b71b24-5036-4167-9c06-e5bfad964f31) -- by 0 seconds Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 277 bytes to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' Nov 27 05:12:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta' Nov 27 05:12:13 localhost systemd[1]: var-lib-containers-storage-overlay-911c17bfaa27c89bad88a1b506a75e32be279f16ba1fd41802de2e8dd0528637-merged.mount: Deactivated successfully. Nov 27 05:12:13 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c2ce19a4dd53a293717e04154ee22c23557207dfa395c356130ffdb4d2250215-userdata-shm.mount: Deactivated successfully. Nov 27 05:12:13 localhost systemd[1]: run-netns-qdhcp\x2d38135c0c\x2d47a1\x2d4390\x2d8c4d\x2d35ca5410096a.mount: Deactivated successfully. Nov 27 05:12:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:13.785 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:12:14 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:14 localhost ovn_controller[156436]: 2025-11-27T10:12:14Z|00330|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:12:14 localhost nova_compute[284026]: 2025-11-27 10:12:14.340 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:12:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:12:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:12:15 localhost systemd[1]: tmp-crun.eFF4Tu.mount: Deactivated successfully. Nov 27 05:12:15 localhost podman[329083]: 2025-11-27 10:12:15.010046658 +0000 UTC m=+0.093370649 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:12:15 localhost podman[329084]: 2025-11-27 10:12:15.028784622 +0000 UTC m=+0.106305907 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, com.redhat.component=ubi9-minimal-container, distribution-scope=public, managed_by=edpm_ansible, vendor=Red Hat, Inc., architecture=x86_64, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, vcs-type=git, release=1755695350, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, url=https://catalog.redhat.com/en/search?searchType=containers, maintainer=Red Hat, Inc., config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:12:15 localhost podman[329083]: 2025-11-27 10:12:15.029937102 +0000 UTC m=+0.113261163 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:12:15 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:12:15 localhost podman[329084]: 2025-11-27 10:12:15.11399135 +0000 UTC m=+0.191512595 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vcs-type=git, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., container_name=openstack_network_exporter, distribution-scope=public, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, url=https://catalog.redhat.com/en/search?searchType=containers, release=1755695350, io.openshift.expose-services=, vendor=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, managed_by=edpm_ansible, config_id=edpm, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, version=9.6, maintainer=Red Hat, Inc.) Nov 27 05:12:15 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:12:15 localhost podman[329082]: 2025-11-27 10:12:15.159107672 +0000 UTC m=+0.245026612 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:12:15 localhost podman[329082]: 2025-11-27 10:12:15.174896887 +0000 UTC m=+0.260815817 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.vendor=CentOS, config_id=edpm, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:12:15 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:12:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v519: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 818 B/s rd, 77 KiB/s wr, 12 op/s Nov 27 05:12:15 localhost systemd[1]: tmp-crun.lNpk0K.mount: Deactivated successfully. Nov 27 05:12:16 localhost nova_compute[284026]: 2025-11-27 10:12:16.140 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] copying data from b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.snap/cf40cff8-dc34-46a6-ac29-9b5a499444e5/0f4b7478-d021-44b5-84d5-0fd9f129c3ec' to b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/490efe1b-1989-4e83-b53b-d78c4d1111e6' Nov 27 05:12:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v520: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 74 KiB/s wr, 12 op/s Nov 27 05:12:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:17 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:12:17 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:17 localhost nova_compute[284026]: 2025-11-27 10:12:17.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 274 bytes to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3 Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] untracking 00cd5267-9c84-40d7-9be4-8a116f4f3d38 Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 151 bytes to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta.tmp' to config b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31/.meta' Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] finished clone: (cephfs, None, 24b71b24-5036-4167-9c06-e5bfad964f31) Nov 27 05:12:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f885b020-cfa9-4f5c-9b53-97c1249d1ff5", "format": "json"}]: dispatch Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "format": "json"}]: dispatch Nov 27 05:12:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:12:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e229 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v521: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 74 KiB/s wr, 12 op/s Nov 27 05:12:21 localhost nova_compute[284026]: 2025-11-27 10:12:21.143 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v522: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 1023 B/s rd, 49 KiB/s wr, 9 op/s Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "format": "json"}]: dispatch Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:22 localhost nova_compute[284026]: 2025-11-27 10:12:22.425 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "e7a60627-9c7e-4350-baca-e4831df56001", "format": "json"}]: dispatch Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:e7a60627-9c7e-4350-baca-e4831df56001, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:e7a60627-9c7e-4350-baca-e4831df56001, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:22.549+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e7a60627-9c7e-4350-baca-e4831df56001' of type subvolume Nov 27 05:12:22 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e7a60627-9c7e-4350-baca-e4831df56001' of type subvolume Nov 27 05:12:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "e7a60627-9c7e-4350-baca-e4831df56001", "force": true, "format": "json"}]: dispatch Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/e7a60627-9c7e-4350-baca-e4831df56001'' moved to trashcan Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e7a60627-9c7e-4350-baca-e4831df56001, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "40bb39ae-97f4-464f-b65e-58ae35704c7e", "format": "json"}]: dispatch Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:22 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e230 e230: 6 total, 6 up, 6 in Nov 27 05:12:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e230 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:23 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:12:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:12:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v524: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 1023 B/s rd, 49 KiB/s wr, 9 op/s Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:23 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:23.600 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:12:23Z, description=, device_id=df4f995c-bd54-4cf2-89c8-bd73dbd638e8, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=5f567a14-16e9-46a9-93ce-0b98652709b7, ip_allocation=immediate, mac_address=fa:16:3e:42:0c:8b, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3121, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:12:23Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:12:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "format": "json"}]: dispatch Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:24b71b24-5036-4167-9c06-e5bfad964f31, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "24b71b24-5036-4167-9c06-e5bfad964f31", "force": true, "format": "json"}]: dispatch Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/24b71b24-5036-4167-9c06-e5bfad964f31'' moved to trashcan Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:24b71b24-5036-4167-9c06-e5bfad964f31, vol_name:cephfs) < "" Nov 27 05:12:23 localhost podman[329157]: 2025-11-27 10:12:23.827192366 +0000 UTC m=+0.039343188 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:12:23 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:12:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:12:23 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:12:23 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:23 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:23 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:23 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:12:24 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:24.261 265123 INFO neutron.agent.dhcp.agent [None req-2b93aee1-6b56-4df5-8345-c2ed26bb926f - - - - - -] DHCP configuration for ports {'5f567a14-16e9-46a9-93ce-0b98652709b7'} is completed#033[00m Nov 27 05:12:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:12:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:12:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "5605670a-9a80-4ad0-b347-ce7327ce598b", "format": "json"}]: dispatch Nov 27 05:12:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:25 localhost podman[329178]: 2025-11-27 10:12:25.04245551 +0000 UTC m=+0.136104257 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, tcib_managed=true, config_id=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:12:25 localhost podman[329179]: 2025-11-27 10:12:25.115218774 +0000 UTC m=+0.204330790 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:12:25 localhost podman[329178]: 2025-11-27 10:12:25.118857643 +0000 UTC m=+0.212506410 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, config_id=ovn_controller, managed_by=edpm_ansible) Nov 27 05:12:25 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:12:25 localhost podman[329179]: 2025-11-27 10:12:25.151871209 +0000 UTC m=+0.240983205 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:12:25 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:12:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v525: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 1023 B/s rd, 49 KiB/s wr, 9 op/s Nov 27 05:12:25 localhost nova_compute[284026]: 2025-11-27 10:12:25.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:25 localhost openstack_network_exporter[244641]: ERROR 10:12:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:12:25 localhost openstack_network_exporter[244641]: ERROR 10:12:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:12:25 localhost openstack_network_exporter[244641]: ERROR 10:12:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:12:25 localhost openstack_network_exporter[244641]: ERROR 10:12:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:12:25 localhost openstack_network_exporter[244641]: Nov 27 05:12:25 localhost openstack_network_exporter[244641]: ERROR 10:12:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:12:25 localhost openstack_network_exporter[244641]: Nov 27 05:12:26 localhost nova_compute[284026]: 2025-11-27 10:12:26.181 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "snap_name": "cf40cff8-dc34-46a6-ac29-9b5a499444e5_31c63848-a827-46e6-8fc4-c0f6f32aef03", "force": true, "format": "json"}]: dispatch Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5_31c63848-a827-46e6-8fc4-c0f6f32aef03, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta' Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5_31c63848-a827-46e6-8fc4-c0f6f32aef03, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "snap_name": "cf40cff8-dc34-46a6-ac29-9b5a499444e5", "force": true, "format": "json"}]: dispatch Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta.tmp' to config b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168/.meta' Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:cf40cff8-dc34-46a6-ac29-9b5a499444e5, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e", "format": "json"}]: dispatch Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:26.875+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e' of type subvolume Nov 27 05:12:26 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e' of type subvolume Nov 27 05:12:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e", "force": true, "format": "json"}]: dispatch Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e'' moved to trashcan Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:ad1cb948-3a42-4794-84a0-f7ec4f5c3c9e, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:27 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:12:27 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v526: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 57 KiB/s wr, 9 op/s Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:27 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:27 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:27 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:12:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3 Nov 27 05:12:27 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e231 e231: 6 total, 6 up, 6 in Nov 27 05:12:27 localhost nova_compute[284026]: 2025-11-27 10:12:27.473 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:27 localhost systemd[1]: tmp-crun.UusHc8.mount: Deactivated successfully. Nov 27 05:12:27 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:12:27 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:12:27 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:12:27 localhost podman[329242]: 2025-11-27 10:12:27.522475629 +0000 UTC m=+0.099938606 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:12:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "6d845cf9-4642-43c6-a545-cf69084d4954", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6d845cf9-4642-43c6-a545-cf69084d4954/.meta.tmp' Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6d845cf9-4642-43c6-a545-cf69084d4954/.meta.tmp' to config b'/volumes/_nogroup/6d845cf9-4642-43c6-a545-cf69084d4954/.meta' Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "6d845cf9-4642-43c6-a545-cf69084d4954", "format": "json"}]: dispatch Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:27 localhost nova_compute[284026]: 2025-11-27 10:12:27.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:27 localhost nova_compute[284026]: 2025-11-27 10:12:27.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:27 localhost nova_compute[284026]: 2025-11-27 10:12:27.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:27 localhost nova_compute[284026]: 2025-11-27 10:12:27.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:12:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e231 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f3ffdf04-08c8-4074-abe8-ed5f955ab972", "format": "json"}]: dispatch Nov 27 05:12:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:28 localhost nova_compute[284026]: 2025-11-27 10:12:28.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v528: 177 pgs: 177 active+clean; 196 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 383 B/s rd, 37 KiB/s wr, 5 op/s Nov 27 05:12:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "format": "json"}]: dispatch Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a864f46f-3fc1-44ad-af0b-573d353ca168, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a864f46f-3fc1-44ad-af0b-573d353ca168, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:30.147+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a864f46f-3fc1-44ad-af0b-573d353ca168' of type subvolume Nov 27 05:12:30 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a864f46f-3fc1-44ad-af0b-573d353ca168' of type subvolume Nov 27 05:12:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "a864f46f-3fc1-44ad-af0b-573d353ca168", "force": true, "format": "json"}]: dispatch Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/a864f46f-3fc1-44ad-af0b-573d353ca168'' moved to trashcan Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a864f46f-3fc1-44ad-af0b-573d353ca168, vol_name:cephfs) < "" Nov 27 05:12:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:12:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:12:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:31 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta' Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:31 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "format": "json"}]: dispatch Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:31 localhost nova_compute[284026]: 2025-11-27 10:12:31.204 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v529: 177 pgs: 177 active+clean; 197 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 1.1 KiB/s rd, 93 KiB/s wr, 13 op/s Nov 27 05:12:31 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:31 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:31 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:31 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:12:31 localhost nova_compute[284026]: 2025-11-27 10:12:31.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "08cea4d4-100a-4c9b-95f5-b5356e1afd81", "format": "json"}]: dispatch Nov 27 05:12:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:32 localhost nova_compute[284026]: 2025-11-27 10:12:32.503 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:32 localhost nova_compute[284026]: 2025-11-27 10:12:32.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e232 e232: 6 total, 6 up, 6 in Nov 27 05:12:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e232 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v531: 177 pgs: 177 active+clean; 197 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 1.1 KiB/s rd, 98 KiB/s wr, 13 op/s Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.769 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.769 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.770 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.770 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:12:33 localhost nova_compute[284026]: 2025-11-27 10:12:33.771 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:12:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:34 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:12:34 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3 Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:12:34 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1410908108' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.289 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.519s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.346 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.347 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:12:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "snap_name": "7e9e7704-db15-4743-83ef-4508674d9944", "format": "json"}]: dispatch Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:7e9e7704-db15-4743-83ef-4508674d9944, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.566 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.568 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11098MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.568 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.568 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:12:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:7e9e7704-db15-4743-83ef-4508674d9944, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.849 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.850 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.850 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.898 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.961 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.962 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 05:12:34 localhost nova_compute[284026]: 2025-11-27 10:12:34.977 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.003 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.040 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:12:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v532: 177 pgs: 177 active+clean; 197 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 767 B/s rd, 61 KiB/s wr, 8 op/s Nov 27 05:12:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:12:35 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3669322113' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.547 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.507s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.554 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.578 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.581 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:12:35 localhost nova_compute[284026]: 2025-11-27 10:12:35.581 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.013s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:12:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "08cea4d4-100a-4c9b-95f5-b5356e1afd81_83b3a1c5-e2d1-4f98-a4ff-1bb95b968367", "force": true, "format": "json"}]: dispatch Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81_83b3a1c5-e2d1-4f98-a4ff-1bb95b968367, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81_83b3a1c5-e2d1-4f98-a4ff-1bb95b968367, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "08cea4d4-100a-4c9b-95f5-b5356e1afd81", "force": true, "format": "json"}]: dispatch Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:08cea4d4-100a-4c9b-95f5-b5356e1afd81, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:36 localhost nova_compute[284026]: 2025-11-27 10:12:36.206 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "tenant_id": "a96b61472b724204a5096c0da386dead", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:12:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:37 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:37 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-1256141615 with tenant a96b61472b724204a5096c0da386dead Nov 27 05:12:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:12:37 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v533: 177 pgs: 177 active+clean; 197 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 829 B/s rd, 87 KiB/s wr, 12 op/s Nov 27 05:12:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume authorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, tenant_id:a96b61472b724204a5096c0da386dead, vol_name:cephfs) < "" Nov 27 05:12:37 localhost nova_compute[284026]: 2025-11-27 10:12:37.541 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:37 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:37 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:37 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:12:37 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-1256141615", "caps": ["mds", "allow rw path=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3", "osd", "allow rw pool=manila_data namespace=fsvolumens_96726c9b-2b08-464e-8d07-49b8e936909c", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:12:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:12:37 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:12:37 localhost podman[329309]: 2025-11-27 10:12:37.994791295 +0000 UTC m=+0.084833029 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible) Nov 27 05:12:38 localhost podman[329309]: 2025-11-27 10:12:37.999863192 +0000 UTC m=+0.089904926 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_managed=true, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:12:38 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:12:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e232 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:38 localhost podman[329310]: 2025-11-27 10:12:38.1006732 +0000 UTC m=+0.188078733 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:12:38 localhost podman[329310]: 2025-11-27 10:12:38.11597943 +0000 UTC m=+0.203384953 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) Nov 27 05:12:38 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:12:38 localhost podman[242678]: time="2025-11-27T10:12:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:12:38 localhost podman[242678]: @ - - [27/Nov/2025:10:12:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:12:38 localhost podman[242678]: @ - - [27/Nov/2025:10:12:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19754 "" "Go-http-client/1.1" Nov 27 05:12:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f3ffdf04-08c8-4074-abe8-ed5f955ab972_2377417a-2701-4208-ab4a-0a7e81db9779", "force": true, "format": "json"}]: dispatch Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972_2377417a-2701-4208-ab4a-0a7e81db9779, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972_2377417a-2701-4208-ab4a-0a7e81db9779, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f3ffdf04-08c8-4074-abe8-ed5f955ab972", "force": true, "format": "json"}]: dispatch Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f3ffdf04-08c8-4074-abe8-ed5f955ab972, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v534: 177 pgs: 177 active+clean; 197 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 818 B/s rd, 85 KiB/s wr, 12 op/s Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.582 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.582 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.583 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.698 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.698 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.699 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:12:39 localhost nova_compute[284026]: 2025-11-27 10:12:39.699 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:12:40 localhost nova_compute[284026]: 2025-11-27 10:12:40.373 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:12:40 localhost nova_compute[284026]: 2025-11-27 10:12:40.398 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:12:40 localhost nova_compute[284026]: 2025-11-27 10:12:40.399 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:12:40 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:40 localhost nova_compute[284026]: 2025-11-27 10:12:40.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} v 0) Nov 27 05:12:40 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} v 0) Nov 27 05:12:40 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume deauthorize, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:40 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "auth_id": "tempest-cephx-id-1256141615", "format": "json"}]: dispatch Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-1256141615, client_metadata.root=/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c/f98be812-0337-46f0-af28-4ab592a979a3 Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:12:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-1256141615, format:json, prefix:fs subvolume evict, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-1256141615", "format": "json"} : dispatch Nov 27 05:12:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"} : dispatch Nov 27 05:12:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-1256141615"}]': finished Nov 27 05:12:41 localhost nova_compute[284026]: 2025-11-27 10:12:41.208 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "snap_name": "7e9e7704-db15-4743-83ef-4508674d9944_20544be6-f283-4e3a-b8b6-a79da6a78e84", "force": true, "format": "json"}]: dispatch Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:7e9e7704-db15-4743-83ef-4508674d9944_20544be6-f283-4e3a-b8b6-a79da6a78e84, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta' Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:7e9e7704-db15-4743-83ef-4508674d9944_20544be6-f283-4e3a-b8b6-a79da6a78e84, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "snap_name": "7e9e7704-db15-4743-83ef-4508674d9944", "force": true, "format": "json"}]: dispatch Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:7e9e7704-db15-4743-83ef-4508674d9944, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta.tmp' to config b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2/.meta' Nov 27 05:12:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v535: 177 pgs: 177 active+clean; 198 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 83 KiB/s wr, 11 op/s Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:7e9e7704-db15-4743-83ef-4508674d9944, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "format": "json"}]: dispatch Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "5605670a-9a80-4ad0-b347-ce7327ce598b_8319d160-7de6-46bc-bc2b-5d51013efa2e", "force": true, "format": "json"}]: dispatch Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b_8319d160-7de6-46bc-bc2b-5d51013efa2e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b_8319d160-7de6-46bc-bc2b-5d51013efa2e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "5605670a-9a80-4ad0-b347-ce7327ce598b", "force": true, "format": "json"}]: dispatch Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:5605670a-9a80-4ad0-b347-ce7327ce598b, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:42 localhost nova_compute[284026]: 2025-11-27 10:12:42.544 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e233 e233: 6 total, 6 up, 6 in Nov 27 05:12:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "0b8cb4bd-755b-4aae-acda-68471d9eccea", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e233 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/0b8cb4bd-755b-4aae-acda-68471d9eccea/.meta.tmp' Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/0b8cb4bd-755b-4aae-acda-68471d9eccea/.meta.tmp' to config b'/volumes/_nogroup/0b8cb4bd-755b-4aae-acda-68471d9eccea/.meta' Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "0b8cb4bd-755b-4aae-acda-68471d9eccea", "format": "json"}]: dispatch Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v537: 177 pgs: 177 active+clean; 198 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 83 KiB/s wr, 11 op/s Nov 27 05:12:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:43.580 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:12:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:12:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:12:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:12:44 Nov 27 05:12:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:12:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:12:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'vms', '.mgr', 'manila_data', 'manila_metadata', 'images', 'volumes'] Nov 27 05:12:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "format": "json"}]: dispatch Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:96726c9b-2b08-464e-8d07-49b8e936909c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:96726c9b-2b08-464e-8d07-49b8e936909c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:44.609+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '96726c9b-2b08-464e-8d07-49b8e936909c' of type subvolume Nov 27 05:12:44 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '96726c9b-2b08-464e-8d07-49b8e936909c' of type subvolume Nov 27 05:12:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "96726c9b-2b08-464e-8d07-49b8e936909c", "force": true, "format": "json"}]: dispatch Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/96726c9b-2b08-464e-8d07-49b8e936909c'' moved to trashcan Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:96726c9b-2b08-464e-8d07-49b8e936909c, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "format": "json"}]: dispatch Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:44.655+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '292ec731-725c-4c7f-82b1-9fe8f3b6cfd2' of type subvolume Nov 27 05:12:44 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '292ec731-725c-4c7f-82b1-9fe8f3b6cfd2' of type subvolume Nov 27 05:12:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "292ec731-725c-4c7f-82b1-9fe8f3b6cfd2", "force": true, "format": "json"}]: dispatch Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/292ec731-725c-4c7f-82b1-9fe8f3b6cfd2'' moved to trashcan Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:292ec731-725c-4c7f-82b1-9fe8f3b6cfd2, vol_name:cephfs) < "" Nov 27 05:12:44 localhost nova_compute[284026]: 2025-11-27 10:12:44.741 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:44 localhost nova_compute[284026]: 2025-11-27 10:12:44.741 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 05:12:44 localhost nova_compute[284026]: 2025-11-27 10:12:44.763 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.9084135957565606e-06 of space, bias 1.0, pg target 0.00037977430555555556 quantized to 32 (current 32) Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:12:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0004149436418202122 of space, bias 4.0, pg target 0.3302951388888889 quantized to 16 (current 16) Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:12:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:12:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "snap_name": "1c150f4d-4fbc-4942-a79a-08a3b53a4dcc", "format": "json"}]: dispatch Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v538: 177 pgs: 177 active+clean; 198 MiB data, 1.0 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 83 KiB/s wr, 11 op/s Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:12:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "40bb39ae-97f4-464f-b65e-58ae35704c7e_2b2cf405-f2eb-4cb8-b8f2-ae69b24f44c1", "force": true, "format": "json"}]: dispatch Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e_2b2cf405-f2eb-4cb8-b8f2-ae69b24f44c1, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e_2b2cf405-f2eb-4cb8-b8f2-ae69b24f44c1, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:45 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "40bb39ae-97f4-464f-b65e-58ae35704c7e", "force": true, "format": "json"}]: dispatch Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:45 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:40bb39ae-97f4-464f-b65e-58ae35704c7e, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:12:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:12:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:12:46 localhost podman[329348]: 2025-11-27 10:12:46.006812677 +0000 UTC m=+0.088533360 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.tags=minimal rhel9, release=1755695350, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, name=ubi9-minimal, com.redhat.component=ubi9-minimal-container, io.openshift.expose-services=, maintainer=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, vendor=Red Hat, Inc., architecture=x86_64, build-date=2025-08-20T13:12:41, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, managed_by=edpm_ansible, vcs-type=git, io.buildah.version=1.33.7, distribution-scope=public, config_id=edpm) Nov 27 05:12:46 localhost podman[329347]: 2025-11-27 10:12:46.054025985 +0000 UTC m=+0.140900676 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:12:46 localhost podman[329347]: 2025-11-27 10:12:46.093949087 +0000 UTC m=+0.180823758 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:12:46 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:12:46 localhost podman[329346]: 2025-11-27 10:12:46.12045013 +0000 UTC m=+0.207665321 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=edpm, container_name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:12:46 localhost podman[329348]: 2025-11-27 10:12:46.12606621 +0000 UTC m=+0.207786953 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, vendor=Red Hat, Inc., config_id=edpm, vcs-type=git, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.tags=minimal rhel9, com.redhat.component=ubi9-minimal-container, version=9.6, maintainer=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, managed_by=edpm_ansible, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, io.openshift.expose-services=) Nov 27 05:12:46 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:12:46 localhost podman[329346]: 2025-11-27 10:12:46.182444565 +0000 UTC m=+0.269659766 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:12:46 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:12:46 localhost nova_compute[284026]: 2025-11-27 10:12:46.212 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "1771e38a-2a50-45e2-a3f6-8cfeb25d3123", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/1771e38a-2a50-45e2-a3f6-8cfeb25d3123/.meta.tmp' Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/1771e38a-2a50-45e2-a3f6-8cfeb25d3123/.meta.tmp' to config b'/volumes/_nogroup/1771e38a-2a50-45e2-a3f6-8cfeb25d3123/.meta' Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "1771e38a-2a50-45e2-a3f6-8cfeb25d3123", "format": "json"}]: dispatch Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:47 localhost systemd[1]: tmp-crun.wlVlEU.mount: Deactivated successfully. Nov 27 05:12:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v539: 177 pgs: 177 active+clean; 198 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 123 KiB/s wr, 14 op/s Nov 27 05:12:47 localhost nova_compute[284026]: 2025-11-27 10:12:47.549 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:47 localhost nova_compute[284026]: 2025-11-27 10:12:47.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:12:47 localhost nova_compute[284026]: 2025-11-27 10:12:47.734 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 05:12:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "6d845cf9-4642-43c6-a545-cf69084d4954", "format": "json"}]: dispatch Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:6d845cf9-4642-43c6-a545-cf69084d4954, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:6d845cf9-4642-43c6-a545-cf69084d4954, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:47 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:47.779+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '6d845cf9-4642-43c6-a545-cf69084d4954' of type subvolume Nov 27 05:12:47 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '6d845cf9-4642-43c6-a545-cf69084d4954' of type subvolume Nov 27 05:12:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "6d845cf9-4642-43c6-a545-cf69084d4954", "force": true, "format": "json"}]: dispatch Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/6d845cf9-4642-43c6-a545-cf69084d4954'' moved to trashcan Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:6d845cf9-4642-43c6-a545-cf69084d4954, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e234 e234: 6 total, 6 up, 6 in Nov 27 05:12:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e234 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f885b020-cfa9-4f5c-9b53-97c1249d1ff5_0ccc7448-8509-43a9-99d5-1983b87b3b2c", "force": true, "format": "json"}]: dispatch Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5_0ccc7448-8509-43a9-99d5-1983b87b3b2c, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5_0ccc7448-8509-43a9-99d5-1983b87b3b2c, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "snap_name": "f885b020-cfa9-4f5c-9b53-97c1249d1ff5", "force": true, "format": "json"}]: dispatch Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta.tmp' to config b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550/.meta' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f885b020-cfa9-4f5c-9b53-97c1249d1ff5, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:12:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/911165648' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:12:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:12:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/911165648' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:12:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot clone", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "snap_name": "1c150f4d-4fbc-4942-a79a-08a3b53a4dcc", "target_sub_name": "c18243b5-6436-4540-a70a-d19043adfb97", "format": "json"}]: dispatch Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, target_sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 273 bytes to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] tracking-id bb0cd3ba-4c78-481d-b4eb-d7046ae5d531 for path b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 246 bytes to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_clone(format:json, prefix:fs subvolume snapshot clone, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, target_sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] cloning to subvolume path: /volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97 Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] starting clone: (cephfs, None, c18243b5-6436-4540-a70a-d19043adfb97) Nov 27 05:12:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "c18243b5-6436-4540-a70a-d19043adfb97", "format": "json"}]: dispatch Nov 27 05:12:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v541: 177 pgs: 177 active+clean; 198 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 96 KiB/s wr, 9 op/s Nov 27 05:12:51 localhost nova_compute[284026]: 2025-11-27 10:12:51.214 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v542: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 1.3 KiB/s rd, 144 KiB/s wr, 16 op/s Nov 27 05:12:52 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:52.096 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=23, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=22) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:12:52 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:52.097 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 10 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:12:52 localhost nova_compute[284026]: 2025-11-27 10:12:52.113 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] Delayed cloning (cephfs, None, c18243b5-6436-4540-a70a-d19043adfb97) -- by 0 seconds Nov 27 05:12:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 277 bytes to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' Nov 27 05:12:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta' Nov 27 05:12:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "1771e38a-2a50-45e2-a3f6-8cfeb25d3123", "format": "json"}]: dispatch Nov 27 05:12:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:52 localhost nova_compute[284026]: 2025-11-27 10:12:52.550 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e235 e235: 6 total, 6 up, 6 in Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #43. Immutable memtables: 0. Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.819877) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 23] Flushing memtable with next log file: 43 Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238372819932, "job": 23, "event": "flush_started", "num_memtables": 1, "num_entries": 2821, "num_deletes": 260, "total_data_size": 4857415, "memory_usage": 4924272, "flush_reason": "Manual Compaction"} Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 23] Level-0 flush table #44: started Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238372841099, "cf_name": "default", "job": 23, "event": "table_file_creation", "file_number": 44, "file_size": 3155790, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 26145, "largest_seqno": 28961, "table_properties": {"data_size": 3144293, "index_size": 7298, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3141, "raw_key_size": 28286, "raw_average_key_size": 22, "raw_value_size": 3119822, "raw_average_value_size": 2491, "num_data_blocks": 306, "num_entries": 1252, "num_filter_entries": 1252, "num_deletions": 260, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238245, "oldest_key_time": 1764238245, "file_creation_time": 1764238372, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 44, "seqno_to_time_mapping": "N/A"}} Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 23] Flush lasted 21278 microseconds, and 8465 cpu microseconds. Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.841159) [db/flush_job.cc:967] [default] [JOB 23] Level-0 flush table #44: 3155790 bytes OK Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.841184) [db/memtable_list.cc:519] [default] Level-0 commit table #44 started Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.851458) [db/memtable_list.cc:722] [default] Level-0 commit table #44: memtable #1 done Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.851485) EVENT_LOG_v1 {"time_micros": 1764238372851477, "job": 23, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.851510) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 23] Try to delete WAL files size 4844032, prev total WAL file size 4844032, number of live WAL files 2. Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000040.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.852605) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003132353530' seq:72057594037927935, type:22 .. '7061786F73003132383032' seq:0, type:0; will stop at (end) Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 24] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 23 Base level 0, inputs: [44(3081KB)], [42(18MB)] Nov 27 05:12:52 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238372852676, "job": 24, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [44], "files_L6": [42], "score": -1, "input_data_size": 22444714, "oldest_snapshot_seqno": -1} Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 24] Generated table #45: 13723 keys, 21148039 bytes, temperature: kUnknown Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238373033143, "cf_name": "default", "job": 24, "event": "table_file_creation", "file_number": 45, "file_size": 21148039, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 21067998, "index_size": 44612, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 34373, "raw_key_size": 366374, "raw_average_key_size": 26, "raw_value_size": 20833050, "raw_average_value_size": 1518, "num_data_blocks": 1685, "num_entries": 13723, "num_filter_entries": 13723, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238372, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 45, "seqno_to_time_mapping": "N/A"}} Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.033453) [db/compaction/compaction_job.cc:1663] [default] [JOB 24] Compacted 1@0 + 1@6 files to L6 => 21148039 bytes Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.035214) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 124.3 rd, 117.1 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(3.0, 18.4 +0.0 blob) out(20.2 +0.0 blob), read-write-amplify(13.8) write-amplify(6.7) OK, records in: 14270, records dropped: 547 output_compression: NoCompression Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.035231) EVENT_LOG_v1 {"time_micros": 1764238373035223, "job": 24, "event": "compaction_finished", "compaction_time_micros": 180571, "compaction_time_cpu_micros": 55187, "output_level": 6, "num_output_files": 1, "total_output_size": 21148039, "num_input_records": 14270, "num_output_records": 13723, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000044.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238373036095, "job": 24, "event": "table_file_deletion", "file_number": 44} Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000042.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238373038356, "job": 24, "event": "table_file_deletion", "file_number": 42} Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:52.852457) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.038408) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.038415) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.038418) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.038420) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:12:53.038423) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:12:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e235 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v544: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 1.4 KiB/s rd, 151 KiB/s wr, 16 op/s Nov 27 05:12:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v545: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 895 B/s rd, 55 KiB/s wr, 7 op/s Nov 27 05:12:55 localhost openstack_network_exporter[244641]: ERROR 10:12:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:12:55 localhost openstack_network_exporter[244641]: ERROR 10:12:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:12:55 localhost openstack_network_exporter[244641]: ERROR 10:12:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:12:55 localhost openstack_network_exporter[244641]: ERROR 10:12:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:12:55 localhost openstack_network_exporter[244641]: Nov 27 05:12:55 localhost openstack_network_exporter[244641]: ERROR 10:12:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:12:55 localhost openstack_network_exporter[244641]: Nov 27 05:12:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:12:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:12:55 localhost podman[329409]: 2025-11-27 10:12:55.98831737 +0000 UTC m=+0.077003029 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=ovn_controller, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:12:56 localhost podman[329410]: 2025-11-27 10:12:56.047213582 +0000 UTC m=+0.131993226 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:12:56 localhost podman[329410]: 2025-11-27 10:12:56.059143322 +0000 UTC m=+0.143922996 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:12:56 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:12:56 localhost podman[329409]: 2025-11-27 10:12:56.074753722 +0000 UTC m=+0.163439371 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_controller) Nov 27 05:12:56 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.265 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:56 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:56.308 265123 INFO neutron.agent.linux.ip_lib [None req-8f4ba0b9-e162-4543-bd82-711f1c6bd299 - - - - - -] Device tap0b158236-fc cannot be used as it has no MAC address#033[00m Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.334 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:56 localhost kernel: device tap0b158236-fc entered promiscuous mode Nov 27 05:12:56 localhost NetworkManager[5971]: [1764238376.3431] manager: (tap0b158236-fc): new Generic device (/org/freedesktop/NetworkManager/Devices/55) Nov 27 05:12:56 localhost ovn_controller[156436]: 2025-11-27T10:12:56Z|00331|binding|INFO|Claiming lport 0b158236-fc28-465f-964a-45db1b3142e4 for this chassis. Nov 27 05:12:56 localhost ovn_controller[156436]: 2025-11-27T10:12:56Z|00332|binding|INFO|0b158236-fc28-465f-964a-45db1b3142e4: Claiming unknown Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.344 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:56 localhost systemd-udevd[329467]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:12:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:56.358 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.103.0.3/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-4fbe7869-19dd-4f94-9e58-ee9b941c17b9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-4fbe7869-19dd-4f94-9e58-ee9b941c17b9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '50343d8ff77a4435ae1c80cd4e9fe260', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=54eebddc-550c-4e55-b165-529bc006a4ce, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=0b158236-fc28-465f-964a-45db1b3142e4) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:12:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:56.360 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 0b158236-fc28-465f-964a-45db1b3142e4 in datapath 4fbe7869-19dd-4f94-9e58-ee9b941c17b9 bound to our chassis#033[00m Nov 27 05:12:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:56.362 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Port 97384045-2fd4-4a69-a521-ed1812e5a771 IP addresses were not retrieved from the Port_Binding MAC column ['unknown'] _get_port_ips /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:536#033[00m Nov 27 05:12:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:56.363 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 4fbe7869-19dd-4f94-9e58-ee9b941c17b9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:12:56 localhost ovn_metadata_agent[162087]: 2025-11-27 10:12:56.364 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[a2d367a0-9abf-49d4-ae2c-075f2de90980]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost ovn_controller[156436]: 2025-11-27T10:12:56Z|00333|binding|INFO|Setting lport 0b158236-fc28-465f-964a-45db1b3142e4 ovn-installed in OVS Nov 27 05:12:56 localhost ovn_controller[156436]: 2025-11-27T10:12:56Z|00334|binding|INFO|Setting lport 0b158236-fc28-465f-964a-45db1b3142e4 up in Southbound Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.384 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost journal[232028]: ethtool ioctl error on tap0b158236-fc: No such device Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.423 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:56 localhost nova_compute[284026]: 2025-11-27 10:12:56.447 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:57.278+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '1771e38a-2a50-45e2-a3f6-8cfeb25d3123' of type subvolume Nov 27 05:12:57 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '1771e38a-2a50-45e2-a3f6-8cfeb25d3123' of type subvolume Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] copying data from b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.snap/1c150f4d-4fbc-4942-a79a-08a3b53a4dcc/268520eb-f690-4f6e-8b05-2face4503f5a' to b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/ee4bb654-ede3-4db3-817f-eba190e77087' Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "1771e38a-2a50-45e2-a3f6-8cfeb25d3123", "force": true, "format": "json"}]: dispatch Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:57 localhost podman[329538]: Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/1771e38a-2a50-45e2-a3f6-8cfeb25d3123'' moved to trashcan Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:1771e38a-2a50-45e2-a3f6-8cfeb25d3123, vol_name:cephfs) < "" Nov 27 05:12:57 localhost podman[329538]: 2025-11-27 10:12:57.310747513 +0000 UTC m=+0.099435191 container create b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "format": "json"}]: dispatch Nov 27 05:12:57 localhost podman[329538]: 2025-11-27 10:12:57.256065025 +0000 UTC m=+0.044752743 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:f98c2e40-6cd7-4323-8649-c131c7189550, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 274 bytes to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta' Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v546: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 1.1 KiB/s rd, 65 KiB/s wr, 8 op/s Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:f98c2e40-6cd7-4323-8649-c131c7189550, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:57 localhost systemd[1]: Started libpod-conmon-b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224.scope. Nov 27 05:12:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:57.367+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f98c2e40-6cd7-4323-8649-c131c7189550' of type subvolume Nov 27 05:12:57 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f98c2e40-6cd7-4323-8649-c131c7189550' of type subvolume Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "f98c2e40-6cd7-4323-8649-c131c7189550", "force": true, "format": "json"}]: dispatch Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:57 localhost systemd[1]: tmp-crun.CQoTXc.mount: Deactivated successfully. Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.clone_index] untracking bb0cd3ba-4c78-481d-b4eb-d7046ae5d531 Nov 27 05:12:57 localhost systemd[1]: Started libcrun container. Nov 27 05:12:57 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/35142a3ac8d51c661b59ab7bd68987ea53b30d7b0e5be1b588a1e62ee4305d64/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:12:57 localhost podman[329538]: 2025-11-27 10:12:57.414135991 +0000 UTC m=+0.202823679 container init b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125) Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta' Nov 27 05:12:57 localhost podman[329538]: 2025-11-27 10:12:57.427675245 +0000 UTC m=+0.216362933 container start b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true) Nov 27 05:12:57 localhost dnsmasq[329556]: started, version 2.85 cachesize 150 Nov 27 05:12:57 localhost dnsmasq[329556]: DNS service limited to local subnets Nov 27 05:12:57 localhost dnsmasq[329556]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:12:57 localhost dnsmasq[329556]: warning: no upstream servers configured Nov 27 05:12:57 localhost dnsmasq-dhcp[329556]: DHCP, static leases only on 10.103.0.0, lease time 1d Nov 27 05:12:57 localhost dnsmasq[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/addn_hosts - 0 addresses Nov 27 05:12:57 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/host Nov 27 05:12:57 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/opts Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 151 bytes to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta.tmp' to config b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97/.meta' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_cloner] finished clone: (cephfs, None, c18243b5-6436-4540-a70a-d19043adfb97) Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/f98c2e40-6cd7-4323-8649-c131c7189550'' moved to trashcan Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f98c2e40-6cd7-4323-8649-c131c7189550, vol_name:cephfs) < "" Nov 27 05:12:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:57.490 265123 INFO neutron.agent.dhcp.agent [None req-bbe15045-b476-4de1-96fc-6aae2fb17b0e - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:12:55Z, description=, device_id=27bdfd31-639a-4433-a790-6cd65b0d08e2, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=9e5c8051-d874-42b5-82ab-f1a103c35dd2, ip_allocation=immediate, mac_address=fa:16:3e:59:9c:9e, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:12:53Z, description=, dns_domain=, id=4fbe7869-19dd-4f94-9e58-ee9b941c17b9, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-RoutersTest-196633868, port_security_enabled=True, project_id=50343d8ff77a4435ae1c80cd4e9fe260, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=30840, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=3223, status=ACTIVE, subnets=['7fe87834-1ccd-4144-9318-cfd2cb0d5c3a'], tags=[], tenant_id=50343d8ff77a4435ae1c80cd4e9fe260, updated_at=2025-11-27T10:12:54Z, vlan_transparent=None, network_id=4fbe7869-19dd-4f94-9e58-ee9b941c17b9, port_security_enabled=False, project_id=50343d8ff77a4435ae1c80cd4e9fe260, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3228, status=DOWN, tags=[], tenant_id=50343d8ff77a4435ae1c80cd4e9fe260, updated_at=2025-11-27T10:12:55Z on network 4fbe7869-19dd-4f94-9e58-ee9b941c17b9#033[00m Nov 27 05:12:57 localhost nova_compute[284026]: 2025-11-27 10:12:57.553 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:12:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:57.588 265123 INFO neutron.agent.dhcp.agent [None req-c5ae1947-5587-4f0c-ab27-f86bb27e35ed - - - - - -] DHCP configuration for ports {'a4238147-90fb-4d06-b73e-6eb443b91944'} is completed#033[00m Nov 27 05:12:57 localhost nova_compute[284026]: 2025-11-27 10:12:57.613 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta' Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:12:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "format": "json"}]: dispatch Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:12:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:12:57 localhost dnsmasq[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/addn_hosts - 1 addresses Nov 27 05:12:57 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/host Nov 27 05:12:57 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/opts Nov 27 05:12:57 localhost podman[329574]: 2025-11-27 10:12:57.765617642 +0000 UTC m=+0.063701642 container kill b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:12:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e236 e236: 6 total, 6 up, 6 in Nov 27 05:12:57 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:57.939 265123 INFO neutron.agent.dhcp.agent [None req-c14d2a84-aac4-47a4-8c9f-32ac3c5d9372 - - - - - -] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:12:55Z, description=, device_id=27bdfd31-639a-4433-a790-6cd65b0d08e2, device_owner=network:router_interface, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=9e5c8051-d874-42b5-82ab-f1a103c35dd2, ip_allocation=immediate, mac_address=fa:16:3e:59:9c:9e, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T10:12:53Z, description=, dns_domain=, id=4fbe7869-19dd-4f94-9e58-ee9b941c17b9, ipv4_address_scope=None, ipv6_address_scope=None, l2_adjacency=True, mtu=1442, name=tempest-RoutersTest-196633868, port_security_enabled=True, project_id=50343d8ff77a4435ae1c80cd4e9fe260, provider:network_type=geneve, provider:physical_network=None, provider:segmentation_id=30840, qos_policy_id=None, revision_number=2, router:external=False, shared=False, standard_attr_id=3223, status=ACTIVE, subnets=['7fe87834-1ccd-4144-9318-cfd2cb0d5c3a'], tags=[], tenant_id=50343d8ff77a4435ae1c80cd4e9fe260, updated_at=2025-11-27T10:12:54Z, vlan_transparent=None, network_id=4fbe7869-19dd-4f94-9e58-ee9b941c17b9, port_security_enabled=False, project_id=50343d8ff77a4435ae1c80cd4e9fe260, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3228, status=DOWN, tags=[], tenant_id=50343d8ff77a4435ae1c80cd4e9fe260, updated_at=2025-11-27T10:12:55Z on network 4fbe7869-19dd-4f94-9e58-ee9b941c17b9#033[00m Nov 27 05:12:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:58.031 265123 INFO neutron.agent.dhcp.agent [None req-a60646ed-2177-473a-8e8e-7e12c9bcc44c - - - - - -] DHCP configuration for ports {'9e5c8051-d874-42b5-82ab-f1a103c35dd2'} is completed#033[00m Nov 27 05:12:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e236 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:12:58 localhost dnsmasq[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/addn_hosts - 1 addresses Nov 27 05:12:58 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/host Nov 27 05:12:58 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/opts Nov 27 05:12:58 localhost podman[329609]: 2025-11-27 10:12:58.156266946 +0000 UTC m=+0.059174830 container kill b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3) Nov 27 05:12:58 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:12:58.558 265123 INFO neutron.agent.dhcp.agent [None req-3bceea9c-8d53-4c24-9468-b91536face6a - - - - - -] DHCP configuration for ports {'9e5c8051-d874-42b5-82ab-f1a103c35dd2'} is completed#033[00m Nov 27 05:12:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "0b8cb4bd-755b-4aae-acda-68471d9eccea", "format": "json"}]: dispatch Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:12:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:12:59.315+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '0b8cb4bd-755b-4aae-acda-68471d9eccea' of type subvolume Nov 27 05:12:59 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '0b8cb4bd-755b-4aae-acda-68471d9eccea' of type subvolume Nov 27 05:12:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "0b8cb4bd-755b-4aae-acda-68471d9eccea", "force": true, "format": "json"}]: dispatch Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/0b8cb4bd-755b-4aae-acda-68471d9eccea'' moved to trashcan Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:0b8cb4bd-755b-4aae-acda-68471d9eccea, vol_name:cephfs) < "" Nov 27 05:12:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v548: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 383 B/s rd, 21 KiB/s wr, 2 op/s Nov 27 05:12:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "snap_name": "e6dc1c73-ce19-4f03-8a96-704ddcd2f328", "format": "json"}]: dispatch Nov 27 05:12:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:12:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e237 e237: 6 total, 6 up, 6 in Nov 27 05:13:00 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:00 localhost dnsmasq[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/addn_hosts - 0 addresses Nov 27 05:13:00 localhost podman[329646]: 2025-11-27 10:13:00.13616645 +0000 UTC m=+0.064170844 container kill b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:13:00 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/host Nov 27 05:13:00 localhost dnsmasq-dhcp[329556]: read /var/lib/neutron/dhcp/4fbe7869-19dd-4f94-9e58-ee9b941c17b9/opts Nov 27 05:13:00 localhost ovn_controller[156436]: 2025-11-27T10:13:00Z|00335|binding|INFO|Releasing lport 0b158236-fc28-465f-964a-45db1b3142e4 from this chassis (sb_readonly=0) Nov 27 05:13:00 localhost ovn_controller[156436]: 2025-11-27T10:13:00Z|00336|binding|INFO|Setting lport 0b158236-fc28-465f-964a-45db1b3142e4 down in Southbound Nov 27 05:13:00 localhost nova_compute[284026]: 2025-11-27 10:13:00.332 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:00 localhost kernel: device tap0b158236-fc left promiscuous mode Nov 27 05:13:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:00.342 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '10.103.0.3/28', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-4fbe7869-19dd-4f94-9e58-ee9b941c17b9', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-4fbe7869-19dd-4f94-9e58-ee9b941c17b9', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': '50343d8ff77a4435ae1c80cd4e9fe260', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=54eebddc-550c-4e55-b165-529bc006a4ce, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=0b158236-fc28-465f-964a-45db1b3142e4) old=Port_Binding(up=[True], chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:13:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:00.345 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 0b158236-fc28-465f-964a-45db1b3142e4 in datapath 4fbe7869-19dd-4f94-9e58-ee9b941c17b9 unbound from our chassis#033[00m Nov 27 05:13:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:00.347 162092 DEBUG neutron.agent.ovn.metadata.agent [-] No valid VIF ports were found for network 4fbe7869-19dd-4f94-9e58-ee9b941c17b9, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:628#033[00m Nov 27 05:13:00 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:00.348 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[6b46540a-fd0f-49a8-a1f3-73b118300257]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:13:00 localhost nova_compute[284026]: 2025-11-27 10:13:00.351 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:00 localhost dnsmasq[329556]: exiting on receipt of SIGTERM Nov 27 05:13:00 localhost podman[329687]: 2025-11-27 10:13:00.834670484 +0000 UTC m=+0.052452031 container kill b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:13:00 localhost systemd[1]: libpod-b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224.scope: Deactivated successfully. Nov 27 05:13:00 localhost podman[329700]: 2025-11-27 10:13:00.907443318 +0000 UTC m=+0.060932748 container died b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:13:00 localhost podman[329700]: 2025-11-27 10:13:00.941083652 +0000 UTC m=+0.094573072 container cleanup b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:13:00 localhost systemd[1]: libpod-conmon-b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224.scope: Deactivated successfully. Nov 27 05:13:00 localhost podman[329702]: 2025-11-27 10:13:00.990416998 +0000 UTC m=+0.133622281 container remove b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-4fbe7869-19dd-4f94-9e58-ee9b941c17b9, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:13:01 localhost systemd[1]: var-lib-containers-storage-overlay-35142a3ac8d51c661b59ab7bd68987ea53b30d7b0e5be1b588a1e62ee4305d64-merged.mount: Deactivated successfully. Nov 27 05:13:01 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-b44cb9af7d71428dc337ab837fe5cffea405bbfacaab4ee656fad14b31001224-userdata-shm.mount: Deactivated successfully. Nov 27 05:13:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:13:01 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1473495853' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:13:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:13:01 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1473495853' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:13:01 localhost nova_compute[284026]: 2025-11-27 10:13:01.299 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:01.301 265123 INFO neutron.agent.dhcp.agent [None req-05c04b76-dd64-439b-8c65-76a06512bb32 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:13:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:01.302 265123 INFO neutron.agent.dhcp.agent [None req-05c04b76-dd64-439b-8c65-76a06512bb32 - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:13:01 localhost systemd[1]: run-netns-qdhcp\x2d4fbe7869\x2d19dd\x2d4f94\x2d9e58\x2dee9b941c17b9.mount: Deactivated successfully. Nov 27 05:13:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v550: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 18 KiB/s rd, 102 KiB/s wr, 32 op/s Nov 27 05:13:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:01.438 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:13:01 localhost ovn_controller[156436]: 2025-11-27T10:13:01Z|00337|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:13:01 localhost nova_compute[284026]: 2025-11-27 10:13:01.696 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:02 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:02.099 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '23'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:13:02 localhost nova_compute[284026]: 2025-11-27 10:13:02.555 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "779ff7fe-723b-4097-8abf-dfb5ecf9abe7", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/779ff7fe-723b-4097-8abf-dfb5ecf9abe7/.meta.tmp' Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/779ff7fe-723b-4097-8abf-dfb5ecf9abe7/.meta.tmp' to config b'/volumes/_nogroup/779ff7fe-723b-4097-8abf-dfb5ecf9abe7/.meta' Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "779ff7fe-723b-4097-8abf-dfb5ecf9abe7", "format": "json"}]: dispatch Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e238 e238: 6 total, 6 up, 6 in Nov 27 05:13:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e238 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v552: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 24 KiB/s rd, 107 KiB/s wr, 40 op/s Nov 27 05:13:05 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "snap_name": "e6dc1c73-ce19-4f03-8a96-704ddcd2f328_814022c6-e812-4295-80da-6068966ad100", "force": true, "format": "json"}]: dispatch Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328_814022c6-e812-4295-80da-6068966ad100, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328_814022c6-e812-4295-80da-6068966ad100, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "snap_name": "e6dc1c73-ce19-4f03-8a96-704ddcd2f328", "force": true, "format": "json"}]: dispatch Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta.tmp' to config b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d/.meta' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e6dc1c73-ce19-4f03-8a96-704ddcd2f328, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "22c37829-54c2-42cf-9024-b4e5b713a0e2", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #45. Immutable memtables: 2. Nov 27 05:13:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v553: 177 pgs: 177 active+clean; 199 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 20 KiB/s rd, 87 KiB/s wr, 32 op/s Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/22c37829-54c2-42cf-9024-b4e5b713a0e2/.meta.tmp' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/22c37829-54c2-42cf-9024-b4e5b713a0e2/.meta.tmp' to config b'/volumes/_nogroup/22c37829-54c2-42cf-9024-b4e5b713a0e2/.meta' Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "22c37829-54c2-42cf-9024-b4e5b713a0e2", "format": "json"}]: dispatch Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "779ff7fe-723b-4097-8abf-dfb5ecf9abe7", "format": "json"}]: dispatch Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:06.076+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '779ff7fe-723b-4097-8abf-dfb5ecf9abe7' of type subvolume Nov 27 05:13:06 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '779ff7fe-723b-4097-8abf-dfb5ecf9abe7' of type subvolume Nov 27 05:13:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "779ff7fe-723b-4097-8abf-dfb5ecf9abe7", "force": true, "format": "json"}]: dispatch Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/779ff7fe-723b-4097-8abf-dfb5ecf9abe7'' moved to trashcan Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:779ff7fe-723b-4097-8abf-dfb5ecf9abe7, vol_name:cephfs) < "" Nov 27 05:13:06 localhost nova_compute[284026]: 2025-11-27 10:13:06.302 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v554: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 34 KiB/s rd, 114 KiB/s wr, 54 op/s Nov 27 05:13:07 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e239 e239: 6 total, 6 up, 6 in Nov 27 05:13:07 localhost nova_compute[284026]: 2025-11-27 10:13:07.582 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e239 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:08 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "format": "json"}]: dispatch Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:08 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:08.437+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '867ebff1-0123-4bf1-853e-2f8e202cbe3d' of type subvolume Nov 27 05:13:08 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '867ebff1-0123-4bf1-853e-2f8e202cbe3d' of type subvolume Nov 27 05:13:08 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "867ebff1-0123-4bf1-853e-2f8e202cbe3d", "force": true, "format": "json"}]: dispatch Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/867ebff1-0123-4bf1-853e-2f8e202cbe3d'' moved to trashcan Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:867ebff1-0123-4bf1-853e-2f8e202cbe3d, vol_name:cephfs) < "" Nov 27 05:13:08 localhost podman[242678]: time="2025-11-27T10:13:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:13:08 localhost podman[242678]: @ - - [27/Nov/2025:10:13:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:13:08 localhost podman[242678]: @ - - [27/Nov/2025:10:13:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19751 "" "Go-http-client/1.1" Nov 27 05:13:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:13:08 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:13:08 localhost podman[329731]: 2025-11-27 10:13:08.987592048 +0000 UTC m=+0.076477865 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.build-date=20251125, config_id=multipathd) Nov 27 05:13:09 localhost podman[329730]: 2025-11-27 10:13:09.050721485 +0000 UTC m=+0.140112975 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:13:09 localhost podman[329731]: 2025-11-27 10:13:09.074340729 +0000 UTC m=+0.163226536 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd) Nov 27 05:13:09 localhost podman[329730]: 2025-11-27 10:13:09.087900733 +0000 UTC m=+0.177292253 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:13:09 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:13:09 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:13:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v556: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 16 KiB/s rd, 34 KiB/s wr, 23 op/s Nov 27 05:13:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:13:11 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:13:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:13:11 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:13:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:13:11 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 239aee93-c197-4452-b963-23bb75c68420 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:13:11 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 239aee93-c197-4452-b963-23bb75c68420 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:13:11 localhost ceph-mgr[290377]: [progress INFO root] Completed event 239aee93-c197-4452-b963-23bb75c68420 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:13:11 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:13:11 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:13:11 localhost nova_compute[284026]: 2025-11-27 10:13:11.335 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:11 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "22c37829-54c2-42cf-9024-b4e5b713a0e2", "format": "json"}]: dispatch Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v557: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 16 KiB/s rd, 74 KiB/s wr, 28 op/s Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:11 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:11.372+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '22c37829-54c2-42cf-9024-b4e5b713a0e2' of type subvolume Nov 27 05:13:11 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '22c37829-54c2-42cf-9024-b4e5b713a0e2' of type subvolume Nov 27 05:13:11 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "22c37829-54c2-42cf-9024-b4e5b713a0e2", "force": true, "format": "json"}]: dispatch Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/22c37829-54c2-42cf-9024-b4e5b713a0e2'' moved to trashcan Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:22c37829-54c2-42cf-9024-b4e5b713a0e2, vol_name:cephfs) < "" Nov 27 05:13:11 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:13:11 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:13:12 localhost nova_compute[284026]: 2025-11-27 10:13:12.625 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 e240: 6 total, 6 up, 6 in Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #46. Immutable memtables: 0. Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.833169) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 25] Flushing memtable with next log file: 46 Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238392833206, "job": 25, "event": "flush_started", "num_memtables": 1, "num_entries": 614, "num_deletes": 260, "total_data_size": 534508, "memory_usage": 545984, "flush_reason": "Manual Compaction"} Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 25] Level-0 flush table #47: started Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238392838260, "cf_name": "default", "job": 25, "event": "table_file_creation", "file_number": 47, "file_size": 345893, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 28966, "largest_seqno": 29575, "table_properties": {"data_size": 342851, "index_size": 961, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1029, "raw_key_size": 8014, "raw_average_key_size": 19, "raw_value_size": 336235, "raw_average_value_size": 828, "num_data_blocks": 42, "num_entries": 406, "num_filter_entries": 406, "num_deletions": 260, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238372, "oldest_key_time": 1764238372, "file_creation_time": 1764238392, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 47, "seqno_to_time_mapping": "N/A"}} Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 25] Flush lasted 5142 microseconds, and 1882 cpu microseconds. Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.838310) [db/flush_job.cc:967] [default] [JOB 25] Level-0 flush table #47: 345893 bytes OK Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.838331) [db/memtable_list.cc:519] [default] Level-0 commit table #47 started Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.840780) [db/memtable_list.cc:722] [default] Level-0 commit table #47: memtable #1 done Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.840805) EVENT_LOG_v1 {"time_micros": 1764238392840799, "job": 25, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.840824) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 25] Try to delete WAL files size 530898, prev total WAL file size 530898, number of live WAL files 2. Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000043.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.841452) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6C6F676D0034323639' seq:72057594037927935, type:22 .. '6C6F676D0034353231' seq:0, type:0; will stop at (end) Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 26] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 25 Base level 0, inputs: [47(337KB)], [45(20MB)] Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238392841502, "job": 26, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [47], "files_L6": [45], "score": -1, "input_data_size": 21493932, "oldest_snapshot_seqno": -1} Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 26] Generated table #48: 13591 keys, 21044988 bytes, temperature: kUnknown Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238392999179, "cf_name": "default", "job": 26, "event": "table_file_creation", "file_number": 48, "file_size": 21044988, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 20965957, "index_size": 43986, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 33989, "raw_key_size": 364792, "raw_average_key_size": 26, "raw_value_size": 20733310, "raw_average_value_size": 1525, "num_data_blocks": 1651, "num_entries": 13591, "num_filter_entries": 13591, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238392, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 48, "seqno_to_time_mapping": "N/A"}} Nov 27 05:13:12 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.999578) [db/compaction/compaction_job.cc:1663] [default] [JOB 26] Compacted 1@0 + 1@6 files to L6 => 21044988 bytes Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.001356) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 136.2 rd, 133.4 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.3, 20.2 +0.0 blob) out(20.1 +0.0 blob), read-write-amplify(123.0) write-amplify(60.8) OK, records in: 14129, records dropped: 538 output_compression: NoCompression Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.001385) EVENT_LOG_v1 {"time_micros": 1764238393001373, "job": 26, "event": "compaction_finished", "compaction_time_micros": 157784, "compaction_time_cpu_micros": 54837, "output_level": 6, "num_output_files": 1, "total_output_size": 21044988, "num_input_records": 14129, "num_output_records": 13591, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000047.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238393001720, "job": 26, "event": "table_file_deletion", "file_number": 47} Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000045.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238393004722, "job": 26, "event": "table_file_deletion", "file_number": 45} Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:12.841326) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.004812) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.004816) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.004818) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.004820) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:13:13.004822) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:13:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v559: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 17 KiB/s rd, 79 KiB/s wr, 30 op/s Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:14 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:13:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:13:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v560: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 1.4 KiB/s rd, 46 KiB/s wr, 6 op/s Nov 27 05:13:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:13:16 localhost nova_compute[284026]: 2025-11-27 10:13:16.369 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:13:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:13:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:13:17 localhost podman[329858]: 2025-11-27 10:13:17.012877875 +0000 UTC m=+0.091470549 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, vcs-type=git, name=ubi9-minimal, distribution-scope=public, architecture=x86_64, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, io.openshift.tags=minimal rhel9, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, container_name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, maintainer=Red Hat, Inc., config_id=edpm, vendor=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b) Nov 27 05:13:17 localhost podman[329856]: 2025-11-27 10:13:17.064250944 +0000 UTC m=+0.150482542 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_id=edpm, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:13:17 localhost podman[329857]: 2025-11-27 10:13:17.110295132 +0000 UTC m=+0.194398504 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:13:17 localhost podman[329857]: 2025-11-27 10:13:17.123946529 +0000 UTC m=+0.208049841 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:13:17 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:13:17 localhost podman[329856]: 2025-11-27 10:13:17.1813625 +0000 UTC m=+0.267594078 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, tcib_managed=true, config_id=edpm, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:13:17 localhost podman[329858]: 2025-11-27 10:13:17.183251282 +0000 UTC m=+0.261843946 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, version=9.6, architecture=x86_64, managed_by=edpm_ansible, io.openshift.tags=minimal rhel9, vcs-type=git, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, release=1755695350, io.openshift.expose-services=, container_name=openstack_network_exporter, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, build-date=2025-08-20T13:12:41, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9.) Nov 27 05:13:17 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:13:17 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:13:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v561: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 49 KiB/s wr, 22 op/s Nov 27 05:13:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "c18243b5-6436-4540-a70a-d19043adfb97", "format": "json"}]: dispatch Nov 27 05:13:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:17 localhost nova_compute[284026]: 2025-11-27 10:13:17.667 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v562: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 49 KiB/s wr, 22 op/s Nov 27 05:13:20 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:20.130 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:13:19Z, description=, device_id=7dd3de6e-3b18-4028-b392-1dcd26dad7c6, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=b8ab1db8-accf-4978-9350-f812e727f7b0, ip_allocation=immediate, mac_address=fa:16:3e:69:65:6b, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3346, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:13:19Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:13:20 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:13:20 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:13:20 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:13:20 localhost podman[329928]: 2025-11-27 10:13:20.379032887 +0000 UTC m=+0.060968309 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_managed=true) Nov 27 05:13:20 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:20.615 265123 INFO neutron.agent.dhcp.agent [None req-79288319-77cd-4cf3-a235-65b5786ccf3c - - - - - -] DHCP configuration for ports {'b8ab1db8-accf-4978-9350-f812e727f7b0'} is completed#033[00m Nov 27 05:13:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v563: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 24 KiB/s wr, 18 op/s Nov 27 05:13:21 localhost nova_compute[284026]: 2025-11-27 10:13:21.420 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:21 localhost nova_compute[284026]: 2025-11-27 10:13:21.798 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:22 localhost nova_compute[284026]: 2025-11-27 10:13:22.707 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "c18243b5-6436-4540-a70a-d19043adfb97", "format": "json"}]: dispatch Nov 27 05:13:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:13:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:13:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "bef2e8e0-366c-45a7-a9df-e90edff8dfc7", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/bef2e8e0-366c-45a7-a9df-e90edff8dfc7/.meta.tmp' Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/bef2e8e0-366c-45a7-a9df-e90edff8dfc7/.meta.tmp' to config b'/volumes/_nogroup/bef2e8e0-366c-45a7-a9df-e90edff8dfc7/.meta' Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "bef2e8e0-366c-45a7-a9df-e90edff8dfc7", "format": "json"}]: dispatch Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v564: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 23 KiB/s wr, 17 op/s Nov 27 05:13:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "bef2e8e0-366c-45a7-a9df-e90edff8dfc7", "format": "json"}]: dispatch Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:24 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:24.935+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'bef2e8e0-366c-45a7-a9df-e90edff8dfc7' of type subvolume Nov 27 05:13:24 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'bef2e8e0-366c-45a7-a9df-e90edff8dfc7' of type subvolume Nov 27 05:13:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "bef2e8e0-366c-45a7-a9df-e90edff8dfc7", "force": true, "format": "json"}]: dispatch Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/bef2e8e0-366c-45a7-a9df-e90edff8dfc7'' moved to trashcan Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:bef2e8e0-366c-45a7-a9df-e90edff8dfc7, vol_name:cephfs) < "" Nov 27 05:13:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v565: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 10 KiB/s rd, 20 KiB/s wr, 15 op/s Nov 27 05:13:25 localhost openstack_network_exporter[244641]: ERROR 10:13:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:13:25 localhost openstack_network_exporter[244641]: ERROR 10:13:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:13:25 localhost openstack_network_exporter[244641]: ERROR 10:13:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:13:25 localhost openstack_network_exporter[244641]: ERROR 10:13:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:13:25 localhost openstack_network_exporter[244641]: Nov 27 05:13:25 localhost openstack_network_exporter[244641]: ERROR 10:13:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:13:25 localhost openstack_network_exporter[244641]: Nov 27 05:13:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "3fcb7c76-414d-4d80-b3ae-2b30b58f4368", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/3fcb7c76-414d-4d80-b3ae-2b30b58f4368/.meta.tmp' Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/3fcb7c76-414d-4d80-b3ae-2b30b58f4368/.meta.tmp' to config b'/volumes/_nogroup/3fcb7c76-414d-4d80-b3ae-2b30b58f4368/.meta' Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "3fcb7c76-414d-4d80-b3ae-2b30b58f4368", "format": "json"}]: dispatch Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:26 localhost nova_compute[284026]: 2025-11-27 10:13:26.463 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:13:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:13:26 localhost systemd[1]: tmp-crun.6M8VTN.mount: Deactivated successfully. Nov 27 05:13:26 localhost podman[329948]: 2025-11-27 10:13:26.995293514 +0000 UTC m=+0.089296679 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, managed_by=edpm_ansible, org.label-schema.build-date=20251125, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:13:27 localhost podman[329948]: 2025-11-27 10:13:27.039198153 +0000 UTC m=+0.133201308 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0) Nov 27 05:13:27 localhost podman[329949]: 2025-11-27 10:13:27.050963159 +0000 UTC m=+0.140580867 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:13:27 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:13:27 localhost podman[329949]: 2025-11-27 10:13:27.087036698 +0000 UTC m=+0.176654366 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:13:27 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:13:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v566: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 10 KiB/s rd, 38 KiB/s wr, 16 op/s Nov 27 05:13:27 localhost nova_compute[284026]: 2025-11-27 10:13:27.744 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:27 localhost nova_compute[284026]: 2025-11-27 10:13:27.749 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:27 localhost nova_compute[284026]: 2025-11-27 10:13:27.749 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:28 localhost nova_compute[284026]: 2025-11-27 10:13:28.184 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:28 localhost nova_compute[284026]: 2025-11-27 10:13:28.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:28 localhost nova_compute[284026]: 2025-11-27 10:13:28.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:13:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume resize", "vol_name": "cephfs", "sub_name": "3fcb7c76-414d-4d80-b3ae-2b30b58f4368", "new_size": 2147483648, "format": "json"}]: dispatch Nov 27 05:13:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_resize(format:json, new_size:2147483648, prefix:fs subvolume resize, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v567: 177 pgs: 177 active+clean; 200 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 85 B/s rd, 27 KiB/s wr, 1 op/s Nov 27 05:13:29 localhost nova_compute[284026]: 2025-11-27 10:13:29.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:29 localhost nova_compute[284026]: 2025-11-27 10:13:29.763 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:30 localhost nova_compute[284026]: 2025-11-27 10:13:30.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:31 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:31.361 265123 INFO neutron.agent.linux.ip_lib [None req-8415adeb-4131-4e74-9055-1b325d48e7a8 - - - - - -] Device tap6b62062d-44 cannot be used as it has no MAC address#033[00m Nov 27 05:13:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v568: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 39 KiB/s wr, 3 op/s Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.437 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:31 localhost kernel: device tap6b62062d-44 entered promiscuous mode Nov 27 05:13:31 localhost ovn_controller[156436]: 2025-11-27T10:13:31Z|00338|binding|INFO|Claiming lport 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 for this chassis. Nov 27 05:13:31 localhost NetworkManager[5971]: [1764238411.4521] manager: (tap6b62062d-44): new Generic device (/org/freedesktop/NetworkManager/Devices/56) Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.451 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:31 localhost ovn_controller[156436]: 2025-11-27T10:13:31Z|00339|binding|INFO|6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9: Claiming unknown Nov 27 05:13:31 localhost systemd-udevd[330009]: Network interface NamePolicy= disabled on kernel command line. Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.464 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:31.479 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: PortBindingUpdatedEvent(events=('update',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[False], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:1::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-e804535e-30fd-4efd-8bc2-900c13dc50ad', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-e804535e-30fd-4efd-8bc2-900c13dc50ad', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'a61f6db1fa9c4abbb0860c0dd4ba063c', 'neutron:revision_number': '1', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c92c2080-176a-4fcd-842e-f0d44b10aea2, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9) old=Port_Binding(chassis=[]) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:13:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:31.481 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 in datapath e804535e-30fd-4efd-8bc2-900c13dc50ad bound to our chassis#033[00m Nov 27 05:13:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:31.482 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network e804535e-30fd-4efd-8bc2-900c13dc50ad or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:13:31 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:31.483 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[c8cfaf5c-1ccd-40e7-9a9a-05896028f9d4]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost ovn_controller[156436]: 2025-11-27T10:13:31Z|00340|binding|INFO|Setting lport 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 ovn-installed in OVS Nov 27 05:13:31 localhost ovn_controller[156436]: 2025-11-27T10:13:31Z|00341|binding|INFO|Setting lport 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 up in Southbound Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.496 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost journal[232028]: ethtool ioctl error on tap6b62062d-44: No such device Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.548 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:31 localhost nova_compute[284026]: 2025-11-27 10:13:31.585 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "3fcb7c76-414d-4d80-b3ae-2b30b58f4368", "format": "json"}]: dispatch Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:32.470+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '3fcb7c76-414d-4d80-b3ae-2b30b58f4368' of type subvolume Nov 27 05:13:32 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '3fcb7c76-414d-4d80-b3ae-2b30b58f4368' of type subvolume Nov 27 05:13:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "3fcb7c76-414d-4d80-b3ae-2b30b58f4368", "force": true, "format": "json"}]: dispatch Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:32 localhost podman[330080]: Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/3fcb7c76-414d-4d80-b3ae-2b30b58f4368'' moved to trashcan Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:3fcb7c76-414d-4d80-b3ae-2b30b58f4368, vol_name:cephfs) < "" Nov 27 05:13:32 localhost podman[330080]: 2025-11-27 10:13:32.491990808 +0000 UTC m=+0.100546002 container create 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:13:32 localhost podman[330080]: 2025-11-27 10:13:32.440031481 +0000 UTC m=+0.048586705 image pull quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified Nov 27 05:13:32 localhost systemd[1]: Started libpod-conmon-2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff.scope. Nov 27 05:13:32 localhost systemd[1]: Started libcrun container. Nov 27 05:13:32 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/b60e5382b60424bcb34b76212de7ab6ed694c3bf0ef3e9100b75aa8b871e0c7b/merged/var/lib/neutron supports timestamps until 2038 (0x7fffffff) Nov 27 05:13:32 localhost podman[330080]: 2025-11-27 10:13:32.58513795 +0000 UTC m=+0.193693144 container init 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:13:32 localhost podman[330080]: 2025-11-27 10:13:32.596787912 +0000 UTC m=+0.205343116 container start 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:13:32 localhost dnsmasq[330098]: started, version 2.85 cachesize 150 Nov 27 05:13:32 localhost dnsmasq[330098]: DNS service limited to local subnets Nov 27 05:13:32 localhost dnsmasq[330098]: compile time options: IPv6 GNU-getopt DBus no-UBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth cryptohash DNSSEC loop-detect inotify dumpfile Nov 27 05:13:32 localhost dnsmasq[330098]: warning: no upstream servers configured Nov 27 05:13:32 localhost dnsmasq-dhcp[330098]: DHCPv6, static leases only on 2001:db8:1::, lease time 1d Nov 27 05:13:32 localhost dnsmasq[330098]: read /var/lib/neutron/dhcp/e804535e-30fd-4efd-8bc2-900c13dc50ad/addn_hosts - 0 addresses Nov 27 05:13:32 localhost dnsmasq-dhcp[330098]: read /var/lib/neutron/dhcp/e804535e-30fd-4efd-8bc2-900c13dc50ad/host Nov 27 05:13:32 localhost dnsmasq-dhcp[330098]: read /var/lib/neutron/dhcp/e804535e-30fd-4efd-8bc2-900c13dc50ad/opts Nov 27 05:13:32 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:32.703 265123 INFO neutron.agent.dhcp.agent [None req-7be5b8eb-0695-48a7-9edd-f889103f2326 - - - - - -] DHCP configuration for ports {'3b79a652-267c-4c26-8415-27a137014112'} is completed#033[00m Nov 27 05:13:32 localhost nova_compute[284026]: 2025-11-27 10:13:32.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:32 localhost nova_compute[284026]: 2025-11-27 10:13:32.747 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "size": 2147483648, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta' Nov 27 05:13:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "format": "json"}]: dispatch Nov 27 05:13:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v569: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 29 KiB/s wr, 2 op/s Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.730 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.753 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.755 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:13:33 localhost nova_compute[284026]: 2025-11-27 10:13:33.755 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:13:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:13:34 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1653601893' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.218 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.462s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.313 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.314 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.515 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.517 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11089MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.517 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.518 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.591 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.592 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.592 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:13:34 localhost nova_compute[284026]: 2025-11-27 10:13:34.637 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:13:34 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:34.827 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:13:34Z, description=, device_id=b83da3ef-7ee2-4e54-8ed0-9520d98b17d1, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=e2ebee24-4ff6-475b-b353-c945250eaf4c, ip_allocation=immediate, mac_address=fa:16:3e:8b:21:56, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3429, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:13:34Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:13:35 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:13:35 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:13:35 localhost podman[330158]: 2025-11-27 10:13:35.060220135 +0000 UTC m=+0.075488129 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS) Nov 27 05:13:35 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:13:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:13:35 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3420984569' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.086 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.449s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.094 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.109 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.112 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.112 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.594s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:13:35 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:35.289 265123 INFO neutron.agent.dhcp.agent [None req-ce3fa469-6549-4244-80c3-f9e4f7befee0 - - - - - -] DHCP configuration for ports {'e2ebee24-4ff6-475b-b353-c945250eaf4c'} is completed#033[00m Nov 27 05:13:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v570: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 29 KiB/s wr, 2 op/s Nov 27 05:13:35 localhost dnsmasq[330098]: exiting on receipt of SIGTERM Nov 27 05:13:35 localhost podman[330198]: 2025-11-27 10:13:35.698617504 +0000 UTC m=+0.063324462 container kill 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0) Nov 27 05:13:35 localhost systemd[1]: libpod-2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff.scope: Deactivated successfully. Nov 27 05:13:35 localhost podman[330211]: 2025-11-27 10:13:35.777346848 +0000 UTC m=+0.061201254 container died 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:13:35 localhost ovn_controller[156436]: 2025-11-27T10:13:35Z|00342|binding|INFO|Removing iface tap6b62062d-44 ovn-installed in OVS Nov 27 05:13:35 localhost ovn_controller[156436]: 2025-11-27T10:13:35Z|00343|binding|INFO|Removing lport 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 ovn-installed in OVS Nov 27 05:13:35 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:35.806 162092 WARNING neutron.agent.ovn.metadata.agent [-] Removing non-external type port fb020f4f-3873-4a10-a74a-358125b89634 with type ""#033[00m Nov 27 05:13:35 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:35.808 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched DELETE: PortBindingDeletedEvent(events=('delete',), table='Port_Binding', conditions=None, old_conditions=None), priority=20 to row=Port_Binding(mac=['unknown'], port_security=[], type=, nat_addresses=[], virtual_parent=[], up=[True], options={'requested-chassis': 'np0005537446.localdomain'}, parent_port=[], requested_additional_chassis=[], ha_chassis_group=[], external_ids={'neutron:cidrs': '2001:db8:1::2/64', 'neutron:device_id': 'dhcp320cf82e-206d-5e66-8137-75b09d6bb97b-e804535e-30fd-4efd-8bc2-900c13dc50ad', 'neutron:device_owner': 'network:dhcp', 'neutron:mtu': '', 'neutron:network_name': 'neutron-e804535e-30fd-4efd-8bc2-900c13dc50ad', 'neutron:port_capabilities': '', 'neutron:port_name': '', 'neutron:project_id': 'a61f6db1fa9c4abbb0860c0dd4ba063c', 'neutron:revision_number': '3', 'neutron:security_group_ids': '', 'neutron:subnet_pool_addr_scope4': '', 'neutron:subnet_pool_addr_scope6': '', 'neutron:vnic_type': 'normal', 'neutron:host_id': 'np0005537446.localdomain'}, additional_chassis=[], tag=[], additional_encap=[], encap=[], mirror_rules=[], datapath=c92c2080-176a-4fcd-842e-f0d44b10aea2, chassis=[], tunnel_key=2, gateway_chassis=[], requested_chassis=[], logical_port=6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9) old= matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.810 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:35 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:35.811 162092 INFO neutron.agent.ovn.metadata.agent [-] Port 6b62062d-44cd-4e07-b7e1-0ba6e3ba15e9 in datapath e804535e-30fd-4efd-8bc2-900c13dc50ad unbound from our chassis#033[00m Nov 27 05:13:35 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:35.812 162092 DEBUG neutron.agent.ovn.metadata.agent [-] There is no metadata port for network e804535e-30fd-4efd-8bc2-900c13dc50ad or it has no MAC or IP addresses configured, tearing the namespace down if needed _get_provision_params /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:599#033[00m Nov 27 05:13:35 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:35.813 162196 DEBUG oslo.privsep.daemon [-] privsep: reply[70b176e3-cf4b-4aad-9884-f3db200d7a24]: (4, False) _call_back /usr/lib/python3.9/site-packages/oslo_privsep/daemon.py:501#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.814 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:35 localhost podman[330211]: 2025-11-27 10:13:35.827394153 +0000 UTC m=+0.111248509 container cleanup 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:13:35 localhost systemd[1]: libpod-conmon-2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff.scope: Deactivated successfully. Nov 27 05:13:35 localhost podman[330212]: 2025-11-27 10:13:35.912720305 +0000 UTC m=+0.189714837 container remove 2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-e804535e-30fd-4efd-8bc2-900c13dc50ad, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:13:35 localhost kernel: device tap6b62062d-44 left promiscuous mode Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.926 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.938 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:35 localhost nova_compute[284026]: 2025-11-27 10:13:35.955 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:35 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:35.965 265123 INFO neutron.agent.dhcp.agent [None req-17d8ddbb-2bfa-4b30-9a03-6442ebc6a29a - - - - - -] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:13:36 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:13:36.052 265123 INFO neutron.agent.dhcp.agent [-] Network not present, action: clean_devices, action_kwargs: {}#033[00m Nov 27 05:13:36 localhost systemd[1]: tmp-crun.I1M02x.mount: Deactivated successfully. Nov 27 05:13:36 localhost systemd[1]: var-lib-containers-storage-overlay-b60e5382b60424bcb34b76212de7ab6ed694c3bf0ef3e9100b75aa8b871e0c7b-merged.mount: Deactivated successfully. Nov 27 05:13:36 localhost systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2a3f826dc1bc7617cc3a11396f00560e68ce26bd0f7331d7fe861dc1acfdb7ff-userdata-shm.mount: Deactivated successfully. Nov 27 05:13:36 localhost systemd[1]: run-netns-qdhcp\x2de804535e\x2d30fd\x2d4efd\x2d8bc2\x2d900c13dc50ad.mount: Deactivated successfully. Nov 27 05:13:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "snap_name": "3a6bdd69-0a75-4471-bcd8-1f4220680219", "format": "json"}]: dispatch Nov 27 05:13:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:36 localhost ovn_controller[156436]: 2025-11-27T10:13:36Z|00344|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:13:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:36 localhost nova_compute[284026]: 2025-11-27 10:13:36.326 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:36 localhost nova_compute[284026]: 2025-11-27 10:13:36.466 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v571: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 50 KiB/s wr, 4 op/s Nov 27 05:13:37 localhost nova_compute[284026]: 2025-11-27 10:13:37.750 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e240 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "be57aae1-0536-4ac6-80df-a61527df14d2", "size": 2147483648, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/be57aae1-0536-4ac6-80df-a61527df14d2/.meta.tmp' Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/be57aae1-0536-4ac6-80df-a61527df14d2/.meta.tmp' to config b'/volumes/_nogroup/be57aae1-0536-4ac6-80df-a61527df14d2/.meta' Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:38 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "be57aae1-0536-4ac6-80df-a61527df14d2", "format": "json"}]: dispatch Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:38 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:38 localhost podman[242678]: time="2025-11-27T10:13:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:13:38 localhost podman[242678]: @ - - [27/Nov/2025:10:13:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:13:38 localhost podman[242678]: @ - - [27/Nov/2025:10:13:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19750 "" "Go-http-client/1.1" Nov 27 05:13:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v572: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 33 KiB/s wr, 3 op/s Nov 27 05:13:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "snap_name": "3a6bdd69-0a75-4471-bcd8-1f4220680219_2161d7a5-ec8d-4f97-8c11-88d65cfd03d2", "force": true, "format": "json"}]: dispatch Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219_2161d7a5-ec8d-4f97-8c11-88d65cfd03d2, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta' Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219_2161d7a5-ec8d-4f97-8c11-88d65cfd03d2, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "snap_name": "3a6bdd69-0a75-4471-bcd8-1f4220680219", "force": true, "format": "json"}]: dispatch Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta.tmp' to config b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd/.meta' Nov 27 05:13:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3a6bdd69-0a75-4471-bcd8-1f4220680219, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:13:39 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:13:39 localhost systemd[1]: tmp-crun.Aeq5yf.mount: Deactivated successfully. Nov 27 05:13:40 localhost podman[330238]: 2025-11-27 10:13:40.00120778 +0000 UTC m=+0.091799817 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:13:40 localhost podman[330238]: 2025-11-27 10:13:40.007723935 +0000 UTC m=+0.098315992 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.build-date=20251125) Nov 27 05:13:40 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:13:40 localhost podman[330239]: 2025-11-27 10:13:40.093971092 +0000 UTC m=+0.179924864 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=multipathd, managed_by=edpm_ansible) Nov 27 05:13:40 localhost podman[330239]: 2025-11-27 10:13:40.105689737 +0000 UTC m=+0.191643489 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, io.buildah.version=1.41.3) Nov 27 05:13:40 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.114 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.115 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.115 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.221 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.913 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.914 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.914 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:13:40 localhost nova_compute[284026]: 2025-11-27 10:13:40.915 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:13:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v573: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 54 KiB/s wr, 5 op/s Nov 27 05:13:41 localhost nova_compute[284026]: 2025-11-27 10:13:41.501 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:41 localhost nova_compute[284026]: 2025-11-27 10:13:41.535 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:13:41 localhost nova_compute[284026]: 2025-11-27 10:13:41.566 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:13:41 localhost nova_compute[284026]: 2025-11-27 10:13:41.566 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:13:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume resize", "vol_name": "cephfs", "sub_name": "be57aae1-0536-4ac6-80df-a61527df14d2", "new_size": 1073741824, "no_shrink": true, "format": "json"}]: dispatch Nov 27 05:13:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_resize(format:json, new_size:1073741824, no_shrink:True, prefix:fs subvolume resize, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_resize(format:json, new_size:1073741824, no_shrink:True, prefix:fs subvolume resize, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:42 localhost nova_compute[284026]: 2025-11-27 10:13:42.754 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "format": "json"}]: dispatch Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:37b54807-d083-44a9-bd54-7631da8fefdd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:37b54807-d083-44a9-bd54-7631da8fefdd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:42 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:42.938+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '37b54807-d083-44a9-bd54-7631da8fefdd' of type subvolume Nov 27 05:13:42 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '37b54807-d083-44a9-bd54-7631da8fefdd' of type subvolume Nov 27 05:13:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "37b54807-d083-44a9-bd54-7631da8fefdd", "force": true, "format": "json"}]: dispatch Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/37b54807-d083-44a9-bd54-7631da8fefdd'' moved to trashcan Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:37b54807-d083-44a9-bd54-7631da8fefdd, vol_name:cephfs) < "" Nov 27 05:13:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e241 e241: 6 total, 6 up, 6 in Nov 27 05:13:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e241 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v575: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 51 KiB/s wr, 4 op/s Nov 27 05:13:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:13:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:13:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:13:43.582 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:13:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:13:43 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2473341125' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:13:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:13:43 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2473341125' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.142 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.147 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3e1dab52-b74a-4471-a100-12fda0e51454', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.143321', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c13597da-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '2d50248cc82bea20e4081e10dbd23c55d88dd15a33a5d511d2655f6fbfe912fc'}]}, 'timestamp': '2025-11-27 10:13:44.148487', '_unique_id': 'a35e39b9ebc14f8e956d2eec05e560d4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.150 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.151 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 18630000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'd718bc42-f953-49f0-b6de-b278badafbfc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 18630000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:13:44.151481', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': 'c138dc92-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.357230766, 'message_signature': '3e0b7d813c1ee029076aec1ac302512917cdd202b3b8c6789ae773c1b8208e82'}]}, 'timestamp': '2025-11-27 10:13:44.169832', '_unique_id': 'a4c27751e4ea47328ae9d015b409174d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.170 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.172 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.172 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '010603e0-a160-4b08-856d-d042affd8eee', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.172190', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c1394b46-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '0674067f0648c27279e01e5d277c983c7d335930cbdf91b09b14f769b1b652c3'}]}, 'timestamp': '2025-11-27 10:13:44.172744', '_unique_id': 'e39d597a2cf247d7810e6ecc9403745e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.174 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.174 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.184 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.185 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e596d295-c41b-472d-99c5-e541874b96d6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.175110', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c13b3e92-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': '4bd1f3568e4563eef2388b44e356a60a2738f0e1644a454bc503d168910966af'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.175110', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c13b5120-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': 'ba9c5776ed7ec1d17f33c7b6eeb793e2d3f5348d120a8988028801b2ada568a6'}]}, 'timestamp': '2025-11-27 10:13:44.185884', '_unique_id': '39697b56d67d40d5b4d2e61455e4c3b0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.186 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.188 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.188 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.188 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8aa700e6-82c2-447e-993c-7ef7ded31277', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.188143', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c13bb9f8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': '79c88cba678b371b9fff2c5056c71ed4bc1838ad282bf51eae005174dca9c46e'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.188143', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c13bcb8c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': '9ffd554483f80e6518d9b8e3cae5f126101119c4e4e5c0a0b3d5269fedca8e6a'}]}, 'timestamp': '2025-11-27 10:13:44.189017', '_unique_id': '98a9053a33f34d16bac3583fa3c86af0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.189 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.191 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.217 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c14a3a60-221f-48a5-b92c-aa02feb16ba8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.191223', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c140315e-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '4990513b87204b2a096ac2d36b294a279d9d4cd637158ce158a725824064dc8e'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.191223', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c140431a-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': 'b43c4c799cf73de232d9cd995beb0745341ef89fe653bfdf6a6aa545cb35505e'}]}, 'timestamp': '2025-11-27 10:13:44.218296', '_unique_id': 'b30bccd80aad47ddb5fc84c59d1ecc2a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.219 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.220 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.220 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a21e0d08-4c4d-474c-ac66-af9cb3aa98a6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.220765', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c140b4bc-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '1fec77d56d983711673963edc9ab4d34ce70512f611089d967f9ab0f277ebaae'}]}, 'timestamp': '2025-11-27 10:13:44.221230', '_unique_id': 'f84b4289428643f3b21038aeede683dc'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.222 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.223 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9442a778-8c0d-453f-a932-bae2b6aa77fc', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.223386', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c1411bfa-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '0b76f11136a053c36427aa80741ba58fbec4ba5a7a5cefa7c7f6be45b967054d'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.223386', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c1412c1c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '91dd87796c881c4ceb3878895efe598367e8665ce7ff965f8aec23b20a133935'}]}, 'timestamp': '2025-11-27 10:13:44.224254', '_unique_id': 'b2845123c8cb4ee4af66ffe2dbd2a6e7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.226 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '7a3df69f-90a9-451c-a0de-2ad0251379df', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.226395', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c141929c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': 'ad15ae75d9a6f0ee436c63e06f5151dd1166a91e57756857abd6e4823238a200'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.226395', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c141be8e-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '41c37bc5574b61665dc78092e12f778cffcd430f5d41c903eab6a28dd4b76a0b'}]}, 'timestamp': '2025-11-27 10:13:44.228018', '_unique_id': '13171d62954e4343b45f39a3d071d02e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.229 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.230 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.231 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b69ce31a-6923-4bad-af63-296829fcd197', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.230618', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c1423558-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '1f897f7a1d8135b25b76ebf030c073eb7321498f04169c39a99b1666aa06b286'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.230618', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c1424548-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': 'd62826b33f210d2da63d49b7f178e029ff622a62d099d0b3a9bb92238181e46b'}]}, 'timestamp': '2025-11-27 10:13:44.231452', '_unique_id': '3a7cb0648aee479da2caf1a8edb47616'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.232 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.233 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c1b6918d-f858-45b8-938f-cc869cd8e68c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.233840', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c142b3ac-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '3291758d7f22080ad3e840af024cdd67c3492cfc31dfc24c0a18a5203339ae32'}]}, 'timestamp': '2025-11-27 10:13:44.234309', '_unique_id': '3d8bddc44113444ab57c990b447d9704'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.235 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.236 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '259b33d7-6423-450e-9bf9-8a162e98dc9f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.236459', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c1431a72-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '1dc6e6f2e13d79edc5c7e857f535d6f105cbf9a63b17f6fcef7389b507513278'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.236459', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c1432a6c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': 'e3900bfddabe595161ef650c6a00765c25b8ec6ba2012748cbec63ae35771f6f'}]}, 'timestamp': '2025-11-27 10:13:44.237318', '_unique_id': '3c297135c985497c9216d1f839eec51a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.238 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.239 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '35da6b9d-551f-49b3-9f9d-2f291312b72a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.239697', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c14397e0-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': 'b9ece2424075212274d933359c9d0048f9d2cf053180cdc3b56dadd7564cfdc3'}]}, 'timestamp': '2025-11-27 10:13:44.240149', '_unique_id': '6ba48e53ed1f4b2085b858e75e3ab9de'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.241 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.242 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9de2ebb4-1fd9-4bc2-9fe4-3d5707d22df3', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.242244', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c143fb2c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': '25382402907deb26d54c9c27adc314fa40d6ef0bf4059a74be41e3934599a92e'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.242244', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c1440f40-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.379287909, 'message_signature': 'dfa3118354794a3789d7d1028a5599746526d808c8fdfb0692bf666d2bef3589'}]}, 'timestamp': '2025-11-27 10:13:44.243180', '_unique_id': 'd8788a25c7e140199b0d8aaab9b8f7ae'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.244 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.245 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '09ed50c5-fb07-4b81-9e83-40cda98221b6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.245313', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c1447462-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '1c48bcb4d2704d5ba65ecfee7a216aa5cb1c2e2a423e0e9426e1e342ce490dbf'}]}, 'timestamp': '2025-11-27 10:13:44.245797', '_unique_id': '9aa04de130934044b46e892ceba2a992'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '226ab234-94f8-4e01-b822-5403ee5622e6', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.248013', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c144dcb8-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '354a005bc560d2eed01f2f21bcf4ff874119cb64e3c724fd97f06b7bc2bf80f7'}]}, 'timestamp': '2025-11-27 10:13:44.248465', '_unique_id': '0ef267df12b640a9862e0808189e2c91'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.250 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3da78347-1609-49fb-9889-db3ee34581ac', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.250830', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c1454b44-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': 'aafbc7f2b4c783af95503a45fec0f2744aa328ad55653dee286948afd1779b4e'}]}, 'timestamp': '2025-11-27 10:13:44.251300', '_unique_id': '4c478814b08e40c084bec94354b727c3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.252 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b2b642a8-937d-4154-9047-53fa431d302a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.253393', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c145b00c-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': '9f10f8dbb4f2a70e1c74ea48734ed7ae2da665449920eff7190ea51cf19ecd7a'}]}, 'timestamp': '2025-11-27 10:13:44.253875', '_unique_id': '6657672f4e1e43c99dd3b23eb272da16'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.255 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4fba5ebe-7394-4e02-9350-8ea974cd307a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:13:44.255887', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': 'c1460cf0-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.357230766, 'message_signature': '7a4ae71cd8fc10d0193952cd0e6c048e96eb2a36db6e6c77d94ddd9d7419779e'}]}, 'timestamp': '2025-11-27 10:13:44.256156', '_unique_id': 'f2ac8817bec64061a37bb69378fecd13'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.256 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.257 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.257 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.257 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4636ec54-e255-42dc-961e-7de5df871c1a', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:13:44.257544', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': 'c1464dd2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.331377682, 'message_signature': 'fc0021ea440f6c82136a0bc6b7282ad038ab44ff2ec6d4b25e5012c6560b7df2'}]}, 'timestamp': '2025-11-27 10:13:44.257827', '_unique_id': '584f67c8a9b84869ad7fe3f9d4418742'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.258 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '2061726c-ae77-4a2a-8b9e-e129318c2866', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:13:44.259238', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': 'c1468fc2-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': 'b2e4171d503725e1c4e1c55bbd2f1ca46f6260945d7442741ea5a02043df14f9'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:13:44.259238', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': 'c1469ad0-cb79-11f0-8166-fa163e6bfee1', 'monotonic_time': 12654.363175706, 'message_signature': 'dd990b535febdc7b69dd20721900becdd514b832853bd23e06098682224bf267'}]}, 'timestamp': '2025-11-27 10:13:44.259782', '_unique_id': '72b2693de6b349b2ac05367b03089810'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:13:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:13:44.260 12 ERROR oslo_messaging.notify.messaging Nov 27 05:13:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:13:44 Nov 27 05:13:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:13:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:13:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['images', 'backups', 'volumes', 'vms', 'manila_metadata', '.mgr', 'manila_data'] Nov 27 05:13:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:13:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "be57aae1-0536-4ac6-80df-a61527df14d2", "format": "json"}]: dispatch Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:be57aae1-0536-4ac6-80df-a61527df14d2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:be57aae1-0536-4ac6-80df-a61527df14d2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:44.785+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'be57aae1-0536-4ac6-80df-a61527df14d2' of type subvolume Nov 27 05:13:44 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'be57aae1-0536-4ac6-80df-a61527df14d2' of type subvolume Nov 27 05:13:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "be57aae1-0536-4ac6-80df-a61527df14d2", "force": true, "format": "json"}]: dispatch Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/be57aae1-0536-4ac6-80df-a61527df14d2'' moved to trashcan Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:be57aae1-0536-4ac6-80df-a61527df14d2, vol_name:cephfs) < "" Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.9084135957565606e-06 of space, bias 1.0, pg target 0.00037977430555555556 quantized to 32 (current 32) Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:13:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0006559490159128978 of space, bias 4.0, pg target 0.5221354166666666 quantized to 16 (current 16) Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:13:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:13:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v576: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 51 KiB/s wr, 4 op/s Nov 27 05:13:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta' Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "format": "json"}]: dispatch Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:13:46 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1484199214' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:13:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:13:46 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1484199214' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:13:46 localhost nova_compute[284026]: 2025-11-27 10:13:46.504 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v577: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 68 KiB/s wr, 23 op/s Nov 27 05:13:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "f24ec14f-55cf-402f-aba9-f9d5749781df", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f24ec14f-55cf-402f-aba9-f9d5749781df/.meta.tmp' Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f24ec14f-55cf-402f-aba9-f9d5749781df/.meta.tmp' to config b'/volumes/_nogroup/f24ec14f-55cf-402f-aba9-f9d5749781df/.meta' Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "f24ec14f-55cf-402f-aba9-f9d5749781df", "format": "json"}]: dispatch Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:47 localhost nova_compute[284026]: 2025-11-27 10:13:47.756 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:13:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:13:47 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:13:47 localhost podman[330277]: 2025-11-27 10:13:47.998829385 +0000 UTC m=+0.083722011 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, maintainer=Red Hat, Inc., name=ubi9-minimal, container_name=openstack_network_exporter, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, architecture=x86_64, io.openshift.tags=minimal rhel9, vendor=Red Hat, Inc., config_id=edpm, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, version=9.6, io.buildah.version=1.33.7) Nov 27 05:13:48 localhost podman[330275]: 2025-11-27 10:13:48.047587314 +0000 UTC m=+0.140868065 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 05:13:48 localhost podman[330277]: 2025-11-27 10:13:48.066606045 +0000 UTC m=+0.151498671 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, name=ubi9-minimal, config_id=edpm, io.buildah.version=1.33.7, vcs-type=git, build-date=2025-08-20T13:12:41, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., release=1755695350, container_name=openstack_network_exporter, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, url=https://catalog.redhat.com/en/search?searchType=containers, managed_by=edpm_ansible, vendor=Red Hat, Inc., version=9.6, io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:13:48 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:13:48 localhost podman[330275]: 2025-11-27 10:13:48.083496299 +0000 UTC m=+0.176777060 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=edpm, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:13:48 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:13:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e241 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "e9e91a76-39f0-4ad3-afc8-0235742f4bce", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:48 localhost podman[330276]: 2025-11-27 10:13:48.147706394 +0000 UTC m=+0.236343000 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:13:48 localhost podman[330276]: 2025-11-27 10:13:48.159940543 +0000 UTC m=+0.248577229 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/e9e91a76-39f0-4ad3-afc8-0235742f4bce/.meta.tmp' Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/e9e91a76-39f0-4ad3-afc8-0235742f4bce/.meta.tmp' to config b'/volumes/_nogroup/e9e91a76-39f0-4ad3-afc8-0235742f4bce/.meta' Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:48 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:13:48 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "e9e91a76-39f0-4ad3-afc8-0235742f4bce", "format": "json"}]: dispatch Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:48 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v578: 177 pgs: 177 active+clean; 201 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 68 KiB/s wr, 23 op/s Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/.meta.tmp' Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/.meta.tmp' to config b'/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/.meta' Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "format": "json"}]: dispatch Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "snap_name": "e85ffe93-bf29-4da7-b8e5-57bf977b1b10", "format": "json"}]: dispatch Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "7024cac5-5f52-4872-be6c-001ad8dab250", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/7024cac5-5f52-4872-be6c-001ad8dab250/.meta.tmp' Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/7024cac5-5f52-4872-be6c-001ad8dab250/.meta.tmp' to config b'/volumes/_nogroup/7024cac5-5f52-4872-be6c-001ad8dab250/.meta' Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "7024cac5-5f52-4872-be6c-001ad8dab250", "format": "json"}]: dispatch Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "f24ec14f-55cf-402f-aba9-f9d5749781df", "format": "json"}]: dispatch Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:f24ec14f-55cf-402f-aba9-f9d5749781df, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:f24ec14f-55cf-402f-aba9-f9d5749781df, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:51.115+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f24ec14f-55cf-402f-aba9-f9d5749781df' of type subvolume Nov 27 05:13:51 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f24ec14f-55cf-402f-aba9-f9d5749781df' of type subvolume Nov 27 05:13:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "f24ec14f-55cf-402f-aba9-f9d5749781df", "force": true, "format": "json"}]: dispatch Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/f24ec14f-55cf-402f-aba9-f9d5749781df'' moved to trashcan Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f24ec14f-55cf-402f-aba9-f9d5749781df, vol_name:cephfs) < "" Nov 27 05:13:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v579: 177 pgs: 177 active+clean; 202 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 39 KiB/s rd, 87 KiB/s wr, 60 op/s Nov 27 05:13:51 localhost nova_compute[284026]: 2025-11-27 10:13:51.542 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "e9e91a76-39f0-4ad3-afc8-0235742f4bce", "format": "json"}]: dispatch Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:52.537+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e9e91a76-39f0-4ad3-afc8-0235742f4bce' of type subvolume Nov 27 05:13:52 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'e9e91a76-39f0-4ad3-afc8-0235742f4bce' of type subvolume Nov 27 05:13:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "e9e91a76-39f0-4ad3-afc8-0235742f4bce", "force": true, "format": "json"}]: dispatch Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/e9e91a76-39f0-4ad3-afc8-0235742f4bce'' moved to trashcan Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:e9e91a76-39f0-4ad3-afc8-0235742f4bce, vol_name:cephfs) < "" Nov 27 05:13:52 localhost nova_compute[284026]: 2025-11-27 10:13:52.758 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:13:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:13:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:13:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:13:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:13:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:13:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e242 e242: 6 total, 6 up, 6 in Nov 27 05:13:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:13:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:13:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:13:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:13:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:13:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e242 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "snap_name": "e85ffe93-bf29-4da7-b8e5-57bf977b1b10_5b82564e-ef31-4d53-bccd-df23f43608d4", "force": true, "format": "json"}]: dispatch Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10_5b82564e-ef31-4d53-bccd-df23f43608d4, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta' Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10_5b82564e-ef31-4d53-bccd-df23f43608d4, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "snap_name": "e85ffe93-bf29-4da7-b8e5-57bf977b1b10", "force": true, "format": "json"}]: dispatch Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta.tmp' to config b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437/.meta' Nov 27 05:13:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e85ffe93-bf29-4da7-b8e5-57bf977b1b10, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v581: 177 pgs: 177 active+clean; 202 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 39 KiB/s rd, 87 KiB/s wr, 60 op/s Nov 27 05:13:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:13:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2574424321' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:13:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:13:54 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2574424321' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:13:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "7024cac5-5f52-4872-be6c-001ad8dab250", "format": "json"}]: dispatch Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:7024cac5-5f52-4872-be6c-001ad8dab250, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:7024cac5-5f52-4872-be6c-001ad8dab250, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:54 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7024cac5-5f52-4872-be6c-001ad8dab250' of type subvolume Nov 27 05:13:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:54.277+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7024cac5-5f52-4872-be6c-001ad8dab250' of type subvolume Nov 27 05:13:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "7024cac5-5f52-4872-be6c-001ad8dab250", "force": true, "format": "json"}]: dispatch Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/7024cac5-5f52-4872-be6c-001ad8dab250'' moved to trashcan Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7024cac5-5f52-4872-be6c-001ad8dab250, vol_name:cephfs) < "" Nov 27 05:13:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v582: 177 pgs: 177 active+clean; 202 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 39 KiB/s rd, 87 KiB/s wr, 60 op/s Nov 27 05:13:55 localhost openstack_network_exporter[244641]: ERROR 10:13:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:13:55 localhost openstack_network_exporter[244641]: ERROR 10:13:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:13:55 localhost openstack_network_exporter[244641]: ERROR 10:13:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:13:55 localhost openstack_network_exporter[244641]: ERROR 10:13:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:13:55 localhost openstack_network_exporter[244641]: Nov 27 05:13:55 localhost openstack_network_exporter[244641]: ERROR 10:13:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:13:55 localhost openstack_network_exporter[244641]: Nov 27 05:13:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "c18243b5-6436-4540-a70a-d19043adfb97", "format": "json"}]: dispatch Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:c18243b5-6436-4540-a70a-d19043adfb97, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:55 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "c18243b5-6436-4540-a70a-d19043adfb97", "force": true, "format": "json"}]: dispatch Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/c18243b5-6436-4540-a70a-d19043adfb97'' moved to trashcan Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:55 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:c18243b5-6436-4540-a70a-d19043adfb97, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:13:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:13:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:13:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "format": "json"}]: dispatch Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:13:56.504+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a64a1e0a-331c-4cdb-af7b-1916c5da3437' of type subvolume Nov 27 05:13:56 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a64a1e0a-331c-4cdb-af7b-1916c5da3437' of type subvolume Nov 27 05:13:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "a64a1e0a-331c-4cdb-af7b-1916c5da3437", "force": true, "format": "json"}]: dispatch Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/a64a1e0a-331c-4cdb-af7b-1916c5da3437'' moved to trashcan Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:13:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a64a1e0a-331c-4cdb-af7b-1916c5da3437, vol_name:cephfs) < "" Nov 27 05:13:56 localhost nova_compute[284026]: 2025-11-27 10:13:56.545 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:13:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:13:57 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:13:57 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:13:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v583: 177 pgs: 177 active+clean; 202 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 50 KiB/s rd, 116 KiB/s wr, 78 op/s Nov 27 05:13:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "658063cb-d598-4b97-9bf5-dee2b18aacd9", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/658063cb-d598-4b97-9bf5-dee2b18aacd9/.meta.tmp' Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/658063cb-d598-4b97-9bf5-dee2b18aacd9/.meta.tmp' to config b'/volumes/_nogroup/658063cb-d598-4b97-9bf5-dee2b18aacd9/.meta' Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:13:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "658063cb-d598-4b97-9bf5-dee2b18aacd9", "format": "json"}]: dispatch Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:13:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:13:57 localhost ovn_controller[156436]: 2025-11-27T10:13:57Z|00345|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:13:57 localhost nova_compute[284026]: 2025-11-27 10:13:57.784 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:57 localhost nova_compute[284026]: 2025-11-27 10:13:57.821 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:13:57 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:13:57 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:13:57 localhost systemd[1]: tmp-crun.E4SHEv.mount: Deactivated successfully. Nov 27 05:13:57 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:13:57 localhost podman[330357]: 2025-11-27 10:13:57.87498223 +0000 UTC m=+0.068146841 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:13:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:13:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:13:57 localhost podman[330371]: 2025-11-27 10:13:57.992073176 +0000 UTC m=+0.086957697 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:13:58 localhost podman[330371]: 2025-11-27 10:13:58.002967839 +0000 UTC m=+0.097852320 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:13:58 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:13:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e243 e243: 6 total, 6 up, 6 in Nov 27 05:13:58 localhost podman[330370]: 2025-11-27 10:13:58.102621226 +0000 UTC m=+0.200732494 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, managed_by=edpm_ansible, io.buildah.version=1.41.3) Nov 27 05:13:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e243 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:13:58 localhost podman[330370]: 2025-11-27 10:13:58.174198398 +0000 UTC m=+0.272309646 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:13:58 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:13:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "snap_name": "1c150f4d-4fbc-4942-a79a-08a3b53a4dcc_bab7f854-77cc-45ed-bedd-95877fe62088", "force": true, "format": "json"}]: dispatch Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc_bab7f854-77cc-45ed-bedd-95877fe62088, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta' Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc_bab7f854-77cc-45ed-bedd-95877fe62088, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:13:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "snap_name": "1c150f4d-4fbc-4942-a79a-08a3b53a4dcc", "force": true, "format": "json"}]: dispatch Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta.tmp' to config b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5/.meta' Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1c150f4d-4fbc-4942-a79a-08a3b53a4dcc, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:13:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e244 e244: 6 total, 6 up, 6 in Nov 27 05:13:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v586: 177 pgs: 177 active+clean; 202 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 39 KiB/s rd, 109 KiB/s wr, 60 op/s Nov 27 05:13:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:13:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:13:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:13:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:13:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:13:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:13:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:00 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:00 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:00 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e245 e245: 6 total, 6 up, 6 in Nov 27 05:14:00 localhost ovn_controller[156436]: 2025-11-27T10:14:00Z|00346|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:14:00 localhost nova_compute[284026]: 2025-11-27 10:14:00.864 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:00 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:14:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:00 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:00 localhost podman[330443]: 2025-11-27 10:14:00.878621066 +0000 UTC m=+0.073198457 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:14:01 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "658063cb-d598-4b97-9bf5-dee2b18aacd9", "format": "json"}]: dispatch Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:01 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:01.118+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '658063cb-d598-4b97-9bf5-dee2b18aacd9' of type subvolume Nov 27 05:14:01 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '658063cb-d598-4b97-9bf5-dee2b18aacd9' of type subvolume Nov 27 05:14:01 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "658063cb-d598-4b97-9bf5-dee2b18aacd9", "force": true, "format": "json"}]: dispatch Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/658063cb-d598-4b97-9bf5-dee2b18aacd9'' moved to trashcan Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:658063cb-d598-4b97-9bf5-dee2b18aacd9, vol_name:cephfs) < "" Nov 27 05:14:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v588: 177 pgs: 2 active+clean+snaptrim, 3 active+clean+snaptrim_wait, 172 active+clean; 203 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 73 KiB/s rd, 257 KiB/s wr, 121 op/s Nov 27 05:14:01 localhost nova_compute[284026]: 2025-11-27 10:14:01.587 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e246 e246: 6 total, 6 up, 6 in Nov 27 05:14:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "format": "json"}]: dispatch Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:33822946-d187-4481-a7b8-0bbdaadb25d5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:33822946-d187-4481-a7b8-0bbdaadb25d5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:02 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:02.241+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '33822946-d187-4481-a7b8-0bbdaadb25d5' of type subvolume Nov 27 05:14:02 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '33822946-d187-4481-a7b8-0bbdaadb25d5' of type subvolume Nov 27 05:14:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "33822946-d187-4481-a7b8-0bbdaadb25d5", "force": true, "format": "json"}]: dispatch Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/33822946-d187-4481-a7b8-0bbdaadb25d5'' moved to trashcan Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:33822946-d187-4481-a7b8-0bbdaadb25d5, vol_name:cephfs) < "" Nov 27 05:14:02 localhost nova_compute[284026]: 2025-11-27 10:14:02.833 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e247 e247: 6 total, 6 up, 6 in Nov 27 05:14:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:14:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:14:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e247 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v591: 177 pgs: 2 active+clean+snaptrim, 3 active+clean+snaptrim_wait, 172 active+clean; 203 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 43 KiB/s rd, 195 KiB/s wr, 79 op/s Nov 27 05:14:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:14:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e248 e248: 6 total, 6 up, 6 in Nov 27 05:14:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v593: 177 pgs: 2 active+clean+snaptrim, 3 active+clean+snaptrim_wait, 172 active+clean; 203 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 4.8 KiB/s wr, 44 op/s Nov 27 05:14:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:14:05 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2450660194' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:14:05 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:14:05 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2450660194' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:14:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:14:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:06 localhost nova_compute[284026]: 2025-11-27 10:14:06.588 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:06 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:06 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:06 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v594: 177 pgs: 177 active+clean; 204 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 82 KiB/s rd, 100 KiB/s wr, 120 op/s Nov 27 05:14:07 localhost nova_compute[284026]: 2025-11-27 10:14:07.836 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e248 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:08 localhost podman[242678]: time="2025-11-27T10:14:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:14:08 localhost podman[242678]: @ - - [27/Nov/2025:10:14:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:14:08 localhost podman[242678]: @ - - [27/Nov/2025:10:14:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19754 "" "Go-http-client/1.1" Nov 27 05:14:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v595: 177 pgs: 177 active+clean; 204 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 63 KiB/s rd, 77 KiB/s wr, 92 op/s Nov 27 05:14:09 localhost nova_compute[284026]: 2025-11-27 10:14:09.505 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:09.507 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=24, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=23) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:14:09 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:09.508 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:14:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:14:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:14:10 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:14:11 localhost systemd[1]: tmp-crun.AEPJaf.mount: Deactivated successfully. Nov 27 05:14:11 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:11 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:11 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:11 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:14:11 localhost podman[330468]: 2025-11-27 10:14:11.040086276 +0000 UTC m=+0.132102569 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible) Nov 27 05:14:11 localhost podman[330468]: 2025-11-27 10:14:11.050957799 +0000 UTC m=+0.142974122 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, config_id=multipathd, container_name=multipathd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, managed_by=edpm_ansible) Nov 27 05:14:11 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:14:11 localhost podman[330467]: 2025-11-27 10:14:11.014747296 +0000 UTC m=+0.109796720 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.license=GPLv2, tcib_managed=true, config_id=ovn_metadata_agent, io.buildah.version=1.41.3) Nov 27 05:14:11 localhost podman[330467]: 2025-11-27 10:14:11.098252079 +0000 UTC m=+0.193301513 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.vendor=CentOS) Nov 27 05:14:11 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:14:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v596: 177 pgs: 177 active+clean; 204 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 76 KiB/s rd, 121 KiB/s wr, 114 op/s Nov 27 05:14:11 localhost nova_compute[284026]: 2025-11-27 10:14:11.590 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:14:12 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:14:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:14:12 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:14:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:14:12 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 57e681f4-8046-4926-a018-127cffe8b872 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:14:12 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 57e681f4-8046-4926-a018-127cffe8b872 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:14:12 localhost ceph-mgr[290377]: [progress INFO root] Completed event 57e681f4-8046-4926-a018-127cffe8b872 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:14:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:14:12 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:14:12 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:12.510 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '24'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:14:12 localhost nova_compute[284026]: 2025-11-27 10:14:12.882 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e249 e249: 6 total, 6 up, 6 in Nov 27 05:14:13 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:14:13 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:14:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e249 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:14:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v598: 177 pgs: 177 active+clean; 204 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 68 KiB/s rd, 109 KiB/s wr, 103 op/s Nov 27 05:14:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:14.548 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:14:14Z, description=, device_id=5950fc59-9e8e-4972-b88a-8afe1b232d7e, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=4330359e-f407-4cdc-87f3-b60ed92e8f3c, ip_allocation=immediate, mac_address=fa:16:3e:0f:1f:59, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3603, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:14:14Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', ), ('cephfs', )] Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:14:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:14:14 localhost systemd[1]: tmp-crun.C7rvn9.mount: Deactivated successfully. Nov 27 05:14:14 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:14:14 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:14 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:14 localhost podman[330605]: 2025-11-27 10:14:14.801099536 +0000 UTC m=+0.059838059 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:14:14 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:14:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:14:15 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:15.059 265123 INFO neutron.agent.dhcp.agent [None req-45f03c31-65ab-4fcd-baca-10208fd4fc7d - - - - - -] DHCP configuration for ports {'4330359e-f407-4cdc-87f3-b60ed92e8f3c'} is completed#033[00m Nov 27 05:14:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v599: 177 pgs: 177 active+clean; 204 MiB data, 1.1 GiB used, 41 GiB / 42 GiB avail; 65 KiB/s rd, 103 KiB/s wr, 97 op/s Nov 27 05:14:15 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:15.578 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:14:15Z, description=, device_id=6b2ddd8f-22c6-4c6f-853b-71e729d27b4f, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=f07c410c-059e-4384-a733-16f2ce8c3d54, ip_allocation=immediate, mac_address=fa:16:3e:b4:1f:30, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3606, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:14:15Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:14:15 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:14:15 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:15 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:15 localhost systemd[1]: tmp-crun.UBWcSy.mount: Deactivated successfully. Nov 27 05:14:15 localhost podman[330645]: 2025-11-27 10:14:15.822633726 +0000 UTC m=+0.062930351 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:14:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:14:16 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:16.126 265123 INFO neutron.agent.dhcp.agent [None req-5e031075-b198-460b-b781-1336b940140d - - - - - -] DHCP configuration for ports {'f07c410c-059e-4384-a733-16f2ce8c3d54'} is completed#033[00m Nov 27 05:14:16 localhost nova_compute[284026]: 2025-11-27 10:14:16.299 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:16 localhost nova_compute[284026]: 2025-11-27 10:14:16.621 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:14:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:16 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:16 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:16 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:16 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:14:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v600: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 28 KiB/s rd, 85 KiB/s wr, 46 op/s Nov 27 05:14:17 localhost nova_compute[284026]: 2025-11-27 10:14:17.498 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:17 localhost nova_compute[284026]: 2025-11-27 10:14:17.916 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e249 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:14:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:14:18 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:14:19 localhost podman[330668]: 2025-11-27 10:14:19.001475449 +0000 UTC m=+0.086911566 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, vendor=Red Hat, Inc., container_name=openstack_network_exporter, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, url=https://catalog.redhat.com/en/search?searchType=containers, distribution-scope=public, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, version=9.6, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-type=git, maintainer=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., managed_by=edpm_ansible, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, config_id=edpm, release=1755695350, io.openshift.expose-services=, architecture=x86_64) Nov 27 05:14:19 localhost podman[330668]: 2025-11-27 10:14:19.015840865 +0000 UTC m=+0.101276992 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, release=1755695350, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_id=edpm, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, name=ubi9-minimal, build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, architecture=x86_64, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vendor=Red Hat, Inc., container_name=openstack_network_exporter, maintainer=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, version=9.6) Nov 27 05:14:19 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:14:19 localhost podman[330667]: 2025-11-27 10:14:19.059811516 +0000 UTC m=+0.145067658 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:14:19 localhost podman[330667]: 2025-11-27 10:14:19.071826088 +0000 UTC m=+0.157082200 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:14:19 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:14:19 localhost podman[330666]: 2025-11-27 10:14:19.158472796 +0000 UTC m=+0.247455178 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, config_id=edpm, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) Nov 27 05:14:19 localhost podman[330666]: 2025-11-27 10:14:19.173657133 +0000 UTC m=+0.262639505 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, config_id=edpm) Nov 27 05:14:19 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:14:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v601: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 28 KiB/s rd, 85 KiB/s wr, 46 op/s Nov 27 05:14:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:14:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:14:19 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:19 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:19 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v602: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 13 KiB/s rd, 85 KiB/s wr, 25 op/s Nov 27 05:14:21 localhost nova_compute[284026]: 2025-11-27 10:14:21.657 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:22 localhost nova_compute[284026]: 2025-11-27 10:14:22.959 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e249 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:23 localhost nova_compute[284026]: 2025-11-27 10:14:23.210 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:14:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:14:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v603: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 12 KiB/s rd, 81 KiB/s wr, 23 op/s Nov 27 05:14:23 localhost nova_compute[284026]: 2025-11-27 10:14:23.654 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:14:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v604: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 11 KiB/s rd, 71 KiB/s wr, 20 op/s Nov 27 05:14:25 localhost openstack_network_exporter[244641]: ERROR 10:14:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:14:25 localhost openstack_network_exporter[244641]: ERROR 10:14:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:14:25 localhost openstack_network_exporter[244641]: ERROR 10:14:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:14:25 localhost openstack_network_exporter[244641]: ERROR 10:14:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:14:25 localhost openstack_network_exporter[244641]: Nov 27 05:14:25 localhost openstack_network_exporter[244641]: ERROR 10:14:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:14:25 localhost openstack_network_exporter[244641]: Nov 27 05:14:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e250 e250: 6 total, 6 up, 6 in Nov 27 05:14:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:26.279 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:14:26Z, description=, device_id=2aa3c57b-8ccb-45ed-a9b3-b986877724c2, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=543a053a-0117-47cc-9845-1433844ebd04, ip_allocation=immediate, mac_address=fa:16:3e:37:b0:06, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3650, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:14:26Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:14:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:14:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:14:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:26 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:26 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 4 addresses Nov 27 05:14:26 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:26 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:26 localhost podman[330744]: 2025-11-27 10:14:26.528442291 +0000 UTC m=+0.061747769 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:14:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:26 localhost nova_compute[284026]: 2025-11-27 10:14:26.659 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:26 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:14:26.848 265123 INFO neutron.agent.dhcp.agent [None req-f1bb7cf0-22e4-41ed-a0e7-9ba664ed6b47 - - - - - -] DHCP configuration for ports {'543a053a-0117-47cc-9845-1433844ebd04'} is completed#033[00m Nov 27 05:14:27 localhost nova_compute[284026]: 2025-11-27 10:14:27.122 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:27 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:27 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:27 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:27 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v606: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 1.8 KiB/s rd, 63 KiB/s wr, 9 op/s Nov 27 05:14:27 localhost nova_compute[284026]: 2025-11-27 10:14:27.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:28 localhost nova_compute[284026]: 2025-11-27 10:14:28.006 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e250 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:14:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:14:29 localhost podman[330766]: 2025-11-27 10:14:29.056310216 +0000 UTC m=+0.146881497 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:14:29 localhost podman[330766]: 2025-11-27 10:14:29.106918275 +0000 UTC m=+0.197489556 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:14:29 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:14:29 localhost systemd[1]: tmp-crun.1JWshc.mount: Deactivated successfully. Nov 27 05:14:29 localhost podman[330765]: 2025-11-27 10:14:29.21205662 +0000 UTC m=+0.303631488 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:14:29 localhost podman[330765]: 2025-11-27 10:14:29.256850973 +0000 UTC m=+0.348425881 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.vendor=CentOS, config_id=ovn_controller, container_name=ovn_controller, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2) Nov 27 05:14:29 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:14:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v607: 177 pgs: 177 active+clean; 205 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 1.8 KiB/s rd, 63 KiB/s wr, 9 op/s Nov 27 05:14:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "87dc15d6-00bd-4c49-94f8-b264b7fba982", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/87dc15d6-00bd-4c49-94f8-b264b7fba982/.meta.tmp' Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/87dc15d6-00bd-4c49-94f8-b264b7fba982/.meta.tmp' to config b'/volumes/_nogroup/87dc15d6-00bd-4c49-94f8-b264b7fba982/.meta' Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:29 localhost nova_compute[284026]: 2025-11-27 10:14:29.687 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "87dc15d6-00bd-4c49-94f8-b264b7fba982", "format": "json"}]: dispatch Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:29 localhost nova_compute[284026]: 2025-11-27 10:14:29.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:29 localhost nova_compute[284026]: 2025-11-27 10:14:29.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:29 localhost nova_compute[284026]: 2025-11-27 10:14:29.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:29 localhost nova_compute[284026]: 2025-11-27 10:14:29.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:14:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:14:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:14:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:14:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:14:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:14:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:14:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:30 localhost nova_compute[284026]: 2025-11-27 10:14:30.734 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v608: 177 pgs: 177 active+clean; 394 MiB data, 1.7 GiB used, 40 GiB / 42 GiB avail; 72 KiB/s rd, 19 MiB/s wr, 115 op/s Nov 27 05:14:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:14:31 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1570903370' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:14:31 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:14:31 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1570903370' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:14:31 localhost nova_compute[284026]: 2025-11-27 10:14:31.705 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:33 localhost nova_compute[284026]: 2025-11-27 10:14:33.010 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "46954a6b-5fb2-41ab-9a85-8f006b4d69e5", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e250 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/46954a6b-5fb2-41ab-9a85-8f006b4d69e5/.meta.tmp' Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/46954a6b-5fb2-41ab-9a85-8f006b4d69e5/.meta.tmp' to config b'/volumes/_nogroup/46954a6b-5fb2-41ab-9a85-8f006b4d69e5/.meta' Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "46954a6b-5fb2-41ab-9a85-8f006b4d69e5", "format": "json"}]: dispatch Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v609: 177 pgs: 177 active+clean; 394 MiB data, 1.7 GiB used, 40 GiB / 42 GiB avail; 72 KiB/s rd, 19 MiB/s wr, 115 op/s Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:14:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:14:33 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:33 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:33 localhost nova_compute[284026]: 2025-11-27 10:14:33.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:33 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.762 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.763 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.763 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:14:34 localhost nova_compute[284026]: 2025-11-27 10:14:34.763 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:14:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e251 e251: 6 total, 6 up, 6 in Nov 27 05:14:35 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:14:35 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/716108350' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.292 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.528s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.373 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.373 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:14:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v611: 177 pgs: 177 active+clean; 394 MiB data, 1.7 GiB used, 40 GiB / 42 GiB avail; 77 KiB/s rd, 21 MiB/s wr, 120 op/s Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.632 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.634 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11052MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.634 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.635 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.737 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.738 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.738 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:14:35 localhost nova_compute[284026]: 2025-11-27 10:14:35.797 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:14:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e252 e252: 6 total, 6 up, 6 in Nov 27 05:14:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:14:36 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2246349988' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.303 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.506s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.311 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.332 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.334 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.335 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.700s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:14:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:36 localhost nova_compute[284026]: 2025-11-27 10:14:36.750 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:14:36 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:14:36 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "5f1563d0-a71b-4144-b5d8-c551a09aaf88", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:37 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:37 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:37 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/5f1563d0-a71b-4144-b5d8-c551a09aaf88/.meta.tmp' Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/5f1563d0-a71b-4144-b5d8-c551a09aaf88/.meta.tmp' to config b'/volumes/_nogroup/5f1563d0-a71b-4144-b5d8-c551a09aaf88/.meta' Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "5f1563d0-a71b-4144-b5d8-c551a09aaf88", "format": "json"}]: dispatch Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v613: 177 pgs: 177 active+clean; 846 MiB data, 2.9 GiB used, 39 GiB / 42 GiB avail; 182 KiB/s rd, 80 MiB/s wr, 300 op/s Nov 27 05:14:38 localhost nova_compute[284026]: 2025-11-27 10:14:38.049 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e252 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:38 localhost podman[242678]: time="2025-11-27T10:14:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:14:38 localhost podman[242678]: @ - - [27/Nov/2025:10:14:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:14:38 localhost podman[242678]: @ - - [27/Nov/2025:10:14:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19750 "" "Go-http-client/1.1" Nov 27 05:14:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v614: 177 pgs: 177 active+clean; 846 MiB data, 2.9 GiB used, 39 GiB / 42 GiB avail; 93 KiB/s rd, 57 MiB/s wr, 162 op/s Nov 27 05:14:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:14:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:14:39 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:39 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:40 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "5eb7208b-1b96-4f2e-905d-2b2608d69fbd", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/5eb7208b-1b96-4f2e-905d-2b2608d69fbd/.meta.tmp' Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/5eb7208b-1b96-4f2e-905d-2b2608d69fbd/.meta.tmp' to config b'/volumes/_nogroup/5eb7208b-1b96-4f2e-905d-2b2608d69fbd/.meta' Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:40 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "5eb7208b-1b96-4f2e-905d-2b2608d69fbd", "format": "json"}]: dispatch Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v615: 177 pgs: 177 active+clean; 1.2 GiB data, 4.1 GiB used, 38 GiB / 42 GiB avail; 198 KiB/s rd, 107 MiB/s wr, 336 op/s Nov 27 05:14:41 localhost nova_compute[284026]: 2025-11-27 10:14:41.789 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:41 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:14:41 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:41 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:41 localhost podman[330875]: 2025-11-27 10:14:41.814627912 +0000 UTC m=+0.063974408 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:14:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:14:41 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:14:41 localhost ovn_controller[156436]: 2025-11-27T10:14:41Z|00347|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:14:41 localhost nova_compute[284026]: 2025-11-27 10:14:41.898 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:41 localhost podman[330889]: 2025-11-27 10:14:41.941326016 +0000 UTC m=+0.094574681 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:14:41 localhost podman[330889]: 2025-11-27 10:14:41.978606398 +0000 UTC m=+0.131855073 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=ovn_metadata_agent, managed_by=edpm_ansible) Nov 27 05:14:41 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:14:42 localhost podman[330890]: 2025-11-27 10:14:41.999809988 +0000 UTC m=+0.150345260 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, managed_by=edpm_ansible) Nov 27 05:14:42 localhost podman[330890]: 2025-11-27 10:14:42.012066676 +0000 UTC m=+0.162601988 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, tcib_managed=true, container_name=multipathd, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, config_id=multipathd) Nov 27 05:14:42 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.336 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.336 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.337 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.439 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.439 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.440 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:14:42 localhost nova_compute[284026]: 2025-11-27 10:14:42.440 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:14:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e253 e253: 6 total, 6 up, 6 in Nov 27 05:14:43 localhost nova_compute[284026]: 2025-11-27 10:14:43.100 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e253 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:43 localhost nova_compute[284026]: 2025-11-27 10:14:43.140 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:14:43 localhost nova_compute[284026]: 2025-11-27 10:14:43.163 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:14:43 localhost nova_compute[284026]: 2025-11-27 10:14:43.164 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:14:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:14:43 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:14:43 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v617: 177 pgs: 177 active+clean; 1.2 GiB data, 4.1 GiB used, 38 GiB / 42 GiB avail; 198 KiB/s rd, 107 MiB/s wr, 336 op/s Nov 27 05:14:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:43.581 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:14:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:43.582 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:14:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:14:43.583 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:14:43 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:14:43 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:43 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:14:43 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:14:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "5eb7208b-1b96-4f2e-905d-2b2608d69fbd", "format": "json"}]: dispatch Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:44 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:44.156+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5eb7208b-1b96-4f2e-905d-2b2608d69fbd' of type subvolume Nov 27 05:14:44 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5eb7208b-1b96-4f2e-905d-2b2608d69fbd' of type subvolume Nov 27 05:14:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "5eb7208b-1b96-4f2e-905d-2b2608d69fbd", "force": true, "format": "json"}]: dispatch Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/5eb7208b-1b96-4f2e-905d-2b2608d69fbd'' moved to trashcan Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5eb7208b-1b96-4f2e-905d-2b2608d69fbd, vol_name:cephfs) < "" Nov 27 05:14:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e254 e254: 6 total, 6 up, 6 in Nov 27 05:14:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:14:44 Nov 27 05:14:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:14:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:14:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'manila_data', 'vms', 'volumes', '.mgr', 'images', 'manila_metadata'] Nov 27 05:14:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:14:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014872903289596129 of space, bias 1.0, pg target 0.296962302348936 quantized to 32 (current 32) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 0.07146954390005583 of space, bias 1.0, pg target 14.222439236111109 quantized to 32 (current 32) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.9084135957565606e-06 of space, bias 1.0, pg target 0.00035305651521496377 quantized to 32 (current 32) Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:14:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0010469011725293131 of space, bias 4.0, pg target 0.7747068676716916 quantized to 16 (current 16) Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:14:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:14:45 localhost ovn_controller[156436]: 2025-11-27T10:14:45Z|00348|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:14:45 localhost nova_compute[284026]: 2025-11-27 10:14:45.266 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:45 localhost systemd[1]: tmp-crun.9Ok5Ey.mount: Deactivated successfully. Nov 27 05:14:45 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:14:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:14:45 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:14:45 localhost podman[330950]: 2025-11-27 10:14:45.336136118 +0000 UTC m=+0.064455724 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:14:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v619: 177 pgs: 177 active+clean; 751 MiB data, 3.1 GiB used, 39 GiB / 42 GiB avail; 134 KiB/s rd, 51 MiB/s wr, 229 op/s Nov 27 05:14:45 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e255 e255: 6 total, 6 up, 6 in Nov 27 05:14:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:14:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:46 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:46 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:46 localhost nova_compute[284026]: 2025-11-27 10:14:46.791 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:47 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:47 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:47 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v621: 177 pgs: 177 active+clean; 207 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 51 KiB/s rd, 67 KiB/s wr, 118 op/s Nov 27 05:14:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "5f1563d0-a71b-4144-b5d8-c551a09aaf88", "format": "json"}]: dispatch Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:47 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:47.448+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5f1563d0-a71b-4144-b5d8-c551a09aaf88' of type subvolume Nov 27 05:14:47 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5f1563d0-a71b-4144-b5d8-c551a09aaf88' of type subvolume Nov 27 05:14:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "5f1563d0-a71b-4144-b5d8-c551a09aaf88", "force": true, "format": "json"}]: dispatch Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/5f1563d0-a71b-4144-b5d8-c551a09aaf88'' moved to trashcan Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5f1563d0-a71b-4144-b5d8-c551a09aaf88, vol_name:cephfs) < "" Nov 27 05:14:48 localhost nova_compute[284026]: 2025-11-27 10:14:48.100 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e255 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:14:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2477938750' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:14:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:14:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2477938750' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:14:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e256 e256: 6 total, 6 up, 6 in Nov 27 05:14:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v623: 177 pgs: 177 active+clean; 207 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 51 KiB/s rd, 67 KiB/s wr, 118 op/s Nov 27 05:14:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:49 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:14:49 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:14:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:14:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:49 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:49 localhost podman[330972]: 2025-11-27 10:14:49.993481894 +0000 UTC m=+0.090170543 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, tcib_managed=true, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, container_name=ceilometer_agent_compute) Nov 27 05:14:50 localhost podman[330972]: 2025-11-27 10:14:50.004578892 +0000 UTC m=+0.101267541 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, io.buildah.version=1.41.3, managed_by=edpm_ansible, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:14:50 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:14:50 localhost podman[330974]: 2025-11-27 10:14:50.060184396 +0000 UTC m=+0.148450779 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, container_name=openstack_network_exporter, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, distribution-scope=public, vcs-type=git, release=1755695350, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, com.redhat.component=ubi9-minimal-container, name=ubi9-minimal, architecture=x86_64, vendor=Red Hat, Inc., maintainer=Red Hat, Inc., version=9.6, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, build-date=2025-08-20T13:12:41, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:14:50 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:50 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:50 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:50 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:14:50 localhost podman[330974]: 2025-11-27 10:14:50.078404075 +0000 UTC m=+0.166670428 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, build-date=2025-08-20T13:12:41, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers, distribution-scope=public, version=9.6, container_name=openstack_network_exporter, vcs-type=git, release=1755695350, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., architecture=x86_64, io.openshift.expose-services=, vendor=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_id=edpm, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, com.redhat.component=ubi9-minimal-container) Nov 27 05:14:50 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:14:50 localhost podman[330973]: 2025-11-27 10:14:50.153002409 +0000 UTC m=+0.244905069 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:14:50 localhost podman[330973]: 2025-11-27 10:14:50.162099513 +0000 UTC m=+0.254002173 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:14:50 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:14:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "46954a6b-5fb2-41ab-9a85-8f006b4d69e5", "format": "json"}]: dispatch Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:50 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:50.612+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '46954a6b-5fb2-41ab-9a85-8f006b4d69e5' of type subvolume Nov 27 05:14:50 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '46954a6b-5fb2-41ab-9a85-8f006b4d69e5' of type subvolume Nov 27 05:14:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "46954a6b-5fb2-41ab-9a85-8f006b4d69e5", "force": true, "format": "json"}]: dispatch Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/46954a6b-5fb2-41ab-9a85-8f006b4d69e5'' moved to trashcan Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:46954a6b-5fb2-41ab-9a85-8f006b4d69e5, vol_name:cephfs) < "" Nov 27 05:14:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "b93b868d-ece3-4f2f-945e-7b84b0f92a4c", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:14:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b93b868d-ece3-4f2f-945e-7b84b0f92a4c/.meta.tmp' Nov 27 05:14:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b93b868d-ece3-4f2f-945e-7b84b0f92a4c/.meta.tmp' to config b'/volumes/_nogroup/b93b868d-ece3-4f2f-945e-7b84b0f92a4c/.meta' Nov 27 05:14:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "b93b868d-ece3-4f2f-945e-7b84b0f92a4c", "format": "json"}]: dispatch Nov 27 05:14:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v624: 177 pgs: 177 active+clean; 208 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 78 KiB/s rd, 142 KiB/s wr, 154 op/s Nov 27 05:14:51 localhost nova_compute[284026]: 2025-11-27 10:14:51.794 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e257 e257: 6 total, 6 up, 6 in Nov 27 05:14:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:14:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:53 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:53 localhost nova_compute[284026]: 2025-11-27 10:14:53.103 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:14:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e257 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v626: 177 pgs: 177 active+clean; 208 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 33 KiB/s rd, 82 KiB/s wr, 52 op/s Nov 27 05:14:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "87dc15d6-00bd-4c49-94f8-b264b7fba982", "format": "json"}]: dispatch Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:53 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '87dc15d6-00bd-4c49-94f8-b264b7fba982' of type subvolume Nov 27 05:14:53 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:53.944+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '87dc15d6-00bd-4c49-94f8-b264b7fba982' of type subvolume Nov 27 05:14:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "87dc15d6-00bd-4c49-94f8-b264b7fba982", "force": true, "format": "json"}]: dispatch Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/87dc15d6-00bd-4c49-94f8-b264b7fba982'' moved to trashcan Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:87dc15d6-00bd-4c49-94f8-b264b7fba982, vol_name:cephfs) < "" Nov 27 05:14:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "b93b868d-ece3-4f2f-945e-7b84b0f92a4c", "format": "json"}]: dispatch Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:14:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:14:54.815+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b93b868d-ece3-4f2f-945e-7b84b0f92a4c' of type subvolume Nov 27 05:14:54 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b93b868d-ece3-4f2f-945e-7b84b0f92a4c' of type subvolume Nov 27 05:14:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "b93b868d-ece3-4f2f-945e-7b84b0f92a4c", "force": true, "format": "json"}]: dispatch Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/b93b868d-ece3-4f2f-945e-7b84b0f92a4c'' moved to trashcan Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:14:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b93b868d-ece3-4f2f-945e-7b84b0f92a4c, vol_name:cephfs) < "" Nov 27 05:14:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v627: 177 pgs: 177 active+clean; 208 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 49 KiB/s rd, 77 KiB/s wr, 73 op/s Nov 27 05:14:55 localhost openstack_network_exporter[244641]: ERROR 10:14:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:14:55 localhost openstack_network_exporter[244641]: ERROR 10:14:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:14:55 localhost openstack_network_exporter[244641]: ERROR 10:14:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:14:55 localhost openstack_network_exporter[244641]: ERROR 10:14:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:14:55 localhost openstack_network_exporter[244641]: Nov 27 05:14:55 localhost openstack_network_exporter[244641]: ERROR 10:14:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:14:55 localhost openstack_network_exporter[244641]: Nov 27 05:14:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:14:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:14:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:14:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:14:56 localhost nova_compute[284026]: 2025-11-27 10:14:56.796 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:14:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:57 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:14:57 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:14:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e258 e258: 6 total, 6 up, 6 in Nov 27 05:14:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v629: 177 pgs: 177 active+clean; 208 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 76 KiB/s rd, 167 KiB/s wr, 113 op/s Nov 27 05:14:58 localhost nova_compute[284026]: 2025-11-27 10:14:58.109 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:14:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e258 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:14:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v630: 177 pgs: 177 active+clean; 208 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 45 KiB/s rd, 91 KiB/s wr, 65 op/s Nov 27 05:14:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:14:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:14:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:14:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:14:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:14:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:14:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:14:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:14:59 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:14:59 localhost systemd[1]: tmp-crun.1Sg1sF.mount: Deactivated successfully. Nov 27 05:15:00 localhost podman[331039]: 2025-11-27 10:15:00.007446031 +0000 UTC m=+0.097979433 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:15:00 localhost podman[331040]: 2025-11-27 10:15:00.098317842 +0000 UTC m=+0.183750647 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:15:00 localhost podman[331039]: 2025-11-27 10:15:00.119202973 +0000 UTC m=+0.209736335 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:15:00 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:15:00 localhost podman[331040]: 2025-11-27 10:15:00.133792215 +0000 UTC m=+0.219225000 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:15:00 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:15:00 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:00 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:00 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e259 e259: 6 total, 6 up, 6 in Nov 27 05:15:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v632: 177 pgs: 177 active+clean; 209 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 95 KiB/s rd, 161 KiB/s wr, 137 op/s Nov 27 05:15:01 localhost nova_compute[284026]: 2025-11-27 10:15:01.799 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:02 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e260 e260: 6 total, 6 up, 6 in Nov 27 05:15:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:15:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:15:03 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3406605242' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:15:03 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/3406605242' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:15:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:03 localhost nova_compute[284026]: 2025-11-27 10:15:03.110 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e260 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v634: 177 pgs: 177 active+clean; 209 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 66 KiB/s rd, 93 KiB/s wr, 96 op/s Nov 27 05:15:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:03 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:15:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:15:04 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1822616431' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:15:04 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:15:04 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1822616431' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #49. Immutable memtables: 0. Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.058818) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 27] Flushing memtable with next log file: 49 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505058870, "job": 27, "event": "flush_started", "num_memtables": 1, "num_entries": 2428, "num_deletes": 261, "total_data_size": 3593504, "memory_usage": 3689664, "flush_reason": "Manual Compaction"} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 27] Level-0 flush table #50: started Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505072719, "cf_name": "default", "job": 27, "event": "table_file_creation", "file_number": 50, "file_size": 2322152, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 29580, "largest_seqno": 32003, "table_properties": {"data_size": 2312606, "index_size": 5731, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 2757, "raw_key_size": 23927, "raw_average_key_size": 22, "raw_value_size": 2292055, "raw_average_value_size": 2126, "num_data_blocks": 247, "num_entries": 1078, "num_filter_entries": 1078, "num_deletions": 261, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238393, "oldest_key_time": 1764238393, "file_creation_time": 1764238505, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 50, "seqno_to_time_mapping": "N/A"}} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 27] Flush lasted 13952 microseconds, and 6265 cpu microseconds. Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.072768) [db/flush_job.cc:967] [default] [JOB 27] Level-0 flush table #50: 2322152 bytes OK Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.072794) [db/memtable_list.cc:519] [default] Level-0 commit table #50 started Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.079892) [db/memtable_list.cc:722] [default] Level-0 commit table #50: memtable #1 done Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.079914) EVENT_LOG_v1 {"time_micros": 1764238505079908, "job": 27, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.079933) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 27] Try to delete WAL files size 3581935, prev total WAL file size 3581935, number of live WAL files 2. Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000046.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.080946) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003132383031' seq:72057594037927935, type:22 .. '7061786F73003133303533' seq:0, type:0; will stop at (end) Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 28] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 27 Base level 0, inputs: [50(2267KB)], [48(20MB)] Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505080987, "job": 28, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [50], "files_L6": [48], "score": -1, "input_data_size": 23367140, "oldest_snapshot_seqno": -1} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 28] Generated table #51: 14125 keys, 21361859 bytes, temperature: kUnknown Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505249390, "cf_name": "default", "job": 28, "event": "table_file_creation", "file_number": 51, "file_size": 21361859, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 21278479, "index_size": 46991, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 35333, "raw_key_size": 377768, "raw_average_key_size": 26, "raw_value_size": 21035813, "raw_average_value_size": 1489, "num_data_blocks": 1769, "num_entries": 14125, "num_filter_entries": 14125, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238505, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 51, "seqno_to_time_mapping": "N/A"}} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.249700) [db/compaction/compaction_job.cc:1663] [default] [JOB 28] Compacted 1@0 + 1@6 files to L6 => 21361859 bytes Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.263060) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 138.7 rd, 126.8 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(2.2, 20.1 +0.0 blob) out(20.4 +0.0 blob), read-write-amplify(19.3) write-amplify(9.2) OK, records in: 14669, records dropped: 544 output_compression: NoCompression Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.263088) EVENT_LOG_v1 {"time_micros": 1764238505263076, "job": 28, "event": "compaction_finished", "compaction_time_micros": 168479, "compaction_time_cpu_micros": 55764, "output_level": 6, "num_output_files": 1, "total_output_size": 21361859, "num_input_records": 14669, "num_output_records": 14125, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000050.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505263483, "job": 28, "event": "table_file_deletion", "file_number": 50} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000048.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238505266362, "job": 28, "event": "table_file_deletion", "file_number": 48} Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.080830) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.266462) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.266470) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.266473) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.266477) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:15:05.266480) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:15:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v635: 177 pgs: 177 active+clean; 209 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 65 KiB/s rd, 71 KiB/s wr, 91 op/s Nov 27 05:15:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "3d7a3eab-87dc-4836-94be-252e607373a3", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/3d7a3eab-87dc-4836-94be-252e607373a3/.meta.tmp' Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/3d7a3eab-87dc-4836-94be-252e607373a3/.meta.tmp' to config b'/volumes/_nogroup/3d7a3eab-87dc-4836-94be-252e607373a3/.meta' Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "3d7a3eab-87dc-4836-94be-252e607373a3", "format": "json"}]: dispatch Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:06 localhost nova_compute[284026]: 2025-11-27 10:15:06.802 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:06 localhost ovn_controller[156436]: 2025-11-27T10:15:06Z|00349|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:15:06 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:15:06 localhost podman[331105]: 2025-11-27 10:15:06.944171098 +0000 UTC m=+0.050314042 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:15:06 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:15:06 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:15:06 localhost nova_compute[284026]: 2025-11-27 10:15:06.973 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v636: 177 pgs: 177 active+clean; 209 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 82 KiB/s rd, 104 KiB/s wr, 116 op/s Nov 27 05:15:08 localhost nova_compute[284026]: 2025-11-27 10:15:08.112 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e260 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:08 localhost podman[242678]: time="2025-11-27T10:15:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:15:08 localhost podman[242678]: @ - - [27/Nov/2025:10:15:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:15:08 localhost podman[242678]: @ - - [27/Nov/2025:10:15:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19745 "" "Go-http-client/1.1" Nov 27 05:15:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v637: 177 pgs: 177 active+clean; 209 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 31 KiB/s rd, 33 KiB/s wr, 43 op/s Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0/.meta.tmp' Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0/.meta.tmp' to config b'/volumes/_nogroup/33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0/.meta' Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0", "format": "json"}]: dispatch Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:09 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:15:09 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:10 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:10 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:10 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:15:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta' Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "format": "json"}]: dispatch Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v638: 177 pgs: 177 active+clean; 210 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 26 KiB/s rd, 87 KiB/s wr, 40 op/s Nov 27 05:15:11 localhost nova_compute[284026]: 2025-11-27 10:15:11.581 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:11.581 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=25, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=24) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:15:11 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:11.583 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 2 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:15:11 localhost nova_compute[284026]: 2025-11-27 10:15:11.804 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0", "format": "json"}]: dispatch Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:12 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:12.693+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0' of type subvolume Nov 27 05:15:12 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0' of type subvolume Nov 27 05:15:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0", "force": true, "format": "json"}]: dispatch Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:15:12 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0'' moved to trashcan Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:33fe4d45-ac0b-45c8-a2e0-60ece5ab5ce0, vol_name:cephfs) < "" Nov 27 05:15:12 localhost podman[331143]: 2025-11-27 10:15:12.829596684 +0000 UTC m=+0.108839075 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_id=ovn_metadata_agent) Nov 27 05:15:12 localhost podman[331143]: 2025-11-27 10:15:12.871616093 +0000 UTC m=+0.150858494 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, managed_by=edpm_ansible, config_id=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:15:12 localhost podman[331145]: 2025-11-27 10:15:12.873944155 +0000 UTC m=+0.146824895 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:15:12 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:15:12 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:15:12 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:12 localhost podman[331145]: 2025-11-27 10:15:12.953631746 +0000 UTC m=+0.226512476 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:15:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:12 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:12 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:12 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:15:12 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e261 e261: 6 total, 6 up, 6 in Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:13 localhost nova_compute[284026]: 2025-11-27 10:15:13.116 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e261 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:13 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:13 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:13 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:13 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v640: 177 pgs: 177 active+clean; 210 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 26 KiB/s rd, 87 KiB/s wr, 40 op/s Nov 27 05:15:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:13.585 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '25'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:15:13 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:15:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:15:13 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev c6620152-0e99-444d-bbff-340014a02d17 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:15:13 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev c6620152-0e99-444d-bbff-340014a02d17 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:15:13 localhost ceph-mgr[290377]: [progress INFO root] Completed event c6620152-0e99-444d-bbff-340014a02d17 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:15:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:15:13 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:15:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "d7fe4d58-768d-40d1-b6f0-9269ec65788f", "format": "json"}]: dispatch Nov 27 05:15:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:15:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:14 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:15:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:15:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v641: 177 pgs: 177 active+clean; 210 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 14 KiB/s rd, 87 KiB/s wr, 26 op/s Nov 27 05:15:15 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "3d7a3eab-87dc-4836-94be-252e607373a3", "format": "json"}]: dispatch Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:3d7a3eab-87dc-4836-94be-252e607373a3, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:3d7a3eab-87dc-4836-94be-252e607373a3, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:15 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:15.914+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '3d7a3eab-87dc-4836-94be-252e607373a3' of type subvolume Nov 27 05:15:15 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '3d7a3eab-87dc-4836-94be-252e607373a3' of type subvolume Nov 27 05:15:15 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "3d7a3eab-87dc-4836-94be-252e607373a3", "force": true, "format": "json"}]: dispatch Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/3d7a3eab-87dc-4836-94be-252e607373a3'' moved to trashcan Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:3d7a3eab-87dc-4836-94be-252e607373a3, vol_name:cephfs) < "" Nov 27 05:15:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:15:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:15:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:16 localhost nova_compute[284026]: 2025-11-27 10:15:16.861 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:15:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "353e4648-540f-431d-9271-83c71c5d3da1", "size": 4294967296, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:4294967296, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/353e4648-540f-431d-9271-83c71c5d3da1/.meta.tmp' Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/353e4648-540f-431d-9271-83c71c5d3da1/.meta.tmp' to config b'/volumes/_nogroup/353e4648-540f-431d-9271-83c71c5d3da1/.meta' Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:4294967296, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "353e4648-540f-431d-9271-83c71c5d3da1", "format": "json"}]: dispatch Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v642: 177 pgs: 177 active+clean; 210 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 121 KiB/s wr, 10 op/s Nov 27 05:15:17 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "c7d9d6be-44f8-4ae4-9632-b7b78c8e1637", "format": "json"}]: dispatch Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:17 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e261 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:18 localhost nova_compute[284026]: 2025-11-27 10:15:18.151 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolumegroup create", "vol_name": "cephfs", "group_name": "868601a0-b39c-46db-a12c-f699e6e7e85b", "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolumegroup_create(format:json, group_name:868601a0-b39c-46db-a12c-f699e6e7e85b, mode:0755, prefix:fs subvolumegroup create, vol_name:cephfs) < "" Nov 27 05:15:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolumegroup_create(format:json, group_name:868601a0-b39c-46db-a12c-f699e6e7e85b, mode:0755, prefix:fs subvolumegroup create, vol_name:cephfs) < "" Nov 27 05:15:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:15:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v643: 177 pgs: 177 active+clean; 210 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 121 KiB/s wr, 10 op/s Nov 27 05:15:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:19 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:19 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:19 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:20 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "869aec1d-b42d-4873-9569-245e8fb1fa0c", "size": 3221225472, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:3221225472, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/869aec1d-b42d-4873-9569-245e8fb1fa0c/.meta.tmp' Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/869aec1d-b42d-4873-9569-245e8fb1fa0c/.meta.tmp' to config b'/volumes/_nogroup/869aec1d-b42d-4873-9569-245e8fb1fa0c/.meta' Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:3221225472, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:20 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "869aec1d-b42d-4873-9569-245e8fb1fa0c", "format": "json"}]: dispatch Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:15:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:15:20 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:15:21 localhost systemd[1]: tmp-crun.V4j0Xk.mount: Deactivated successfully. Nov 27 05:15:21 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "c7d9d6be-44f8-4ae4-9632-b7b78c8e1637_9e939470-2a1c-40da-b10b-f1ca8317a495", "force": true, "format": "json"}]: dispatch Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637_9e939470-2a1c-40da-b10b-f1ca8317a495, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:21 localhost podman[331251]: 2025-11-27 10:15:21.055180702 +0000 UTC m=+0.134776021 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, name=ubi9-minimal, build-date=2025-08-20T13:12:41, release=1755695350, architecture=x86_64, vcs-type=git, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.buildah.version=1.33.7, maintainer=Red Hat, Inc., vendor=Red Hat, Inc., container_name=openstack_network_exporter, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, managed_by=edpm_ansible, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta' Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637_9e939470-2a1c-40da-b10b-f1ca8317a495, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:21 localhost podman[331250]: 2025-11-27 10:15:21.027718165 +0000 UTC m=+0.108121215 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:15:21 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "c7d9d6be-44f8-4ae4-9632-b7b78c8e1637", "force": true, "format": "json"}]: dispatch Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:21 localhost podman[331251]: 2025-11-27 10:15:21.097829698 +0000 UTC m=+0.177425047 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-type=git, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, version=9.6, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, io.openshift.tags=minimal rhel9, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., io.openshift.expose-services=, name=ubi9-minimal, managed_by=edpm_ansible, release=1755695350, architecture=x86_64, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_id=edpm, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, container_name=openstack_network_exporter) Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta' Nov 27 05:15:21 localhost podman[331250]: 2025-11-27 10:15:21.11207566 +0000 UTC m=+0.192478730 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:15:21 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:15:21 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:15:21 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:c7d9d6be-44f8-4ae4-9632-b7b78c8e1637, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:21 localhost podman[331249]: 2025-11-27 10:15:21.216094944 +0000 UTC m=+0.298685143 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2) Nov 27 05:15:21 localhost podman[331249]: 2025-11-27 10:15:21.227119701 +0000 UTC m=+0.309709890 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:15:21 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:15:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v644: 177 pgs: 177 active+clean; 211 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 614 B/s rd, 135 KiB/s wr, 11 op/s Nov 27 05:15:21 localhost nova_compute[284026]: 2025-11-27 10:15:21.863 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolumegroup rm", "vol_name": "cephfs", "group_name": "868601a0-b39c-46db-a12c-f699e6e7e85b", "force": true, "format": "json"}]: dispatch Nov 27 05:15:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolumegroup_rm(force:True, format:json, group_name:868601a0-b39c-46db-a12c-f699e6e7e85b, prefix:fs subvolumegroup rm, vol_name:cephfs) < "" Nov 27 05:15:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolumegroup_rm(force:True, format:json, group_name:868601a0-b39c-46db-a12c-f699e6e7e85b, prefix:fs subvolumegroup rm, vol_name:cephfs) < "" Nov 27 05:15:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e261 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:23 localhost nova_compute[284026]: 2025-11-27 10:15:23.153 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v645: 177 pgs: 177 active+clean; 211 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 588 B/s rd, 129 KiB/s wr, 10 op/s Nov 27 05:15:24 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:24 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:24 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:15:24 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "353e4648-540f-431d-9271-83c71c5d3da1", "format": "json"}]: dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:353e4648-540f-431d-9271-83c71c5d3da1, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:353e4648-540f-431d-9271-83c71c5d3da1, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:24.493+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '353e4648-540f-431d-9271-83c71c5d3da1' of type subvolume Nov 27 05:15:24 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '353e4648-540f-431d-9271-83c71c5d3da1' of type subvolume Nov 27 05:15:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "353e4648-540f-431d-9271-83c71c5d3da1", "force": true, "format": "json"}]: dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/353e4648-540f-431d-9271-83c71c5d3da1'' moved to trashcan Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:353e4648-540f-431d-9271-83c71c5d3da1, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "d7fe4d58-768d-40d1-b6f0-9269ec65788f_a283d428-8a52-4098-b4f6-4c4eb4f92be8", "force": true, "format": "json"}]: dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f_a283d428-8a52-4098-b4f6-4c4eb4f92be8, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta' Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f_a283d428-8a52-4098-b4f6-4c4eb4f92be8, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "snap_name": "d7fe4d58-768d-40d1-b6f0-9269ec65788f", "force": true, "format": "json"}]: dispatch Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta.tmp' to config b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213/.meta' Nov 27 05:15:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:d7fe4d58-768d-40d1-b6f0-9269ec65788f, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:25 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:25 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:15:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolumegroup create", "vol_name": "cephfs", "group_name": "97da0ccb-1d29-41f7-aaa9-f077b69c7349", "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolumegroup_create(format:json, group_name:97da0ccb-1d29-41f7-aaa9-f077b69c7349, mode:0755, prefix:fs subvolumegroup create, vol_name:cephfs) < "" Nov 27 05:15:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v646: 177 pgs: 177 active+clean; 211 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 767 B/s rd, 113 KiB/s wr, 10 op/s Nov 27 05:15:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolumegroup_create(format:json, group_name:97da0ccb-1d29-41f7-aaa9-f077b69c7349, mode:0755, prefix:fs subvolumegroup create, vol_name:cephfs) < "" Nov 27 05:15:25 localhost openstack_network_exporter[244641]: ERROR 10:15:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:15:25 localhost openstack_network_exporter[244641]: ERROR 10:15:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:15:25 localhost openstack_network_exporter[244641]: ERROR 10:15:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:15:25 localhost openstack_network_exporter[244641]: ERROR 10:15:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:15:25 localhost openstack_network_exporter[244641]: Nov 27 05:15:25 localhost openstack_network_exporter[244641]: ERROR 10:15:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:15:25 localhost openstack_network_exporter[244641]: Nov 27 05:15:25 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:15:25 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:15:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:26 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:26 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:26 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:26 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:26 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:26 localhost nova_compute[284026]: 2025-11-27 10:15:26.866 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "869aec1d-b42d-4873-9569-245e8fb1fa0c", "format": "json"}]: dispatch Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:27.052+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '869aec1d-b42d-4873-9569-245e8fb1fa0c' of type subvolume Nov 27 05:15:27 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '869aec1d-b42d-4873-9569-245e8fb1fa0c' of type subvolume Nov 27 05:15:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "869aec1d-b42d-4873-9569-245e8fb1fa0c", "force": true, "format": "json"}]: dispatch Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/869aec1d-b42d-4873-9569-245e8fb1fa0c'' moved to trashcan Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:869aec1d-b42d-4873-9569-245e8fb1fa0c, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v647: 177 pgs: 177 active+clean; 212 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 177 KiB/s wr, 13 op/s Nov 27 05:15:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "473d9190-3882-48f1-b6b4-b83854651213", "format": "json"}]: dispatch Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:473d9190-3882-48f1-b6b4-b83854651213, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:473d9190-3882-48f1-b6b4-b83854651213, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:27.448+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '473d9190-3882-48f1-b6b4-b83854651213' of type subvolume Nov 27 05:15:27 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '473d9190-3882-48f1-b6b4-b83854651213' of type subvolume Nov 27 05:15:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "473d9190-3882-48f1-b6b4-b83854651213", "force": true, "format": "json"}]: dispatch Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/473d9190-3882-48f1-b6b4-b83854651213'' moved to trashcan Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:473d9190-3882-48f1-b6b4-b83854651213, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e261 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:28 localhost nova_compute[284026]: 2025-11-27 10:15:28.156 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e262 e262: 6 total, 6 up, 6 in Nov 27 05:15:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolumegroup rm", "vol_name": "cephfs", "group_name": "97da0ccb-1d29-41f7-aaa9-f077b69c7349", "force": true, "format": "json"}]: dispatch Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolumegroup_rm(force:True, format:json, group_name:97da0ccb-1d29-41f7-aaa9-f077b69c7349, prefix:fs subvolumegroup rm, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolumegroup_rm(force:True, format:json, group_name:97da0ccb-1d29-41f7-aaa9-f077b69c7349, prefix:fs subvolumegroup rm, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "4bbff30f-e3b8-4d51-834f-f8208de38bd5", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/4bbff30f-e3b8-4d51-834f-f8208de38bd5/.meta.tmp' Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/4bbff30f-e3b8-4d51-834f-f8208de38bd5/.meta.tmp' to config b'/volumes/_nogroup/4bbff30f-e3b8-4d51-834f-f8208de38bd5/.meta' Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "4bbff30f-e3b8-4d51-834f-f8208de38bd5", "format": "json"}]: dispatch Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v649: 177 pgs: 177 active+clean; 212 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 614 B/s rd, 153 KiB/s wr, 12 op/s Nov 27 05:15:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:15:29 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:29 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:15:29 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:29 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:29 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:29 localhost nova_compute[284026]: 2025-11-27 10:15:29.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:29 localhost nova_compute[284026]: 2025-11-27 10:15:29.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:15:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:15:30 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:15:31 localhost podman[331315]: 2025-11-27 10:15:31.006960388 +0000 UTC m=+0.090253445 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:15:31 localhost systemd[1]: tmp-crun.PVmdQu.mount: Deactivated successfully. Nov 27 05:15:31 localhost podman[331316]: 2025-11-27 10:15:31.091454399 +0000 UTC m=+0.167625794 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:15:31 localhost podman[331315]: 2025-11-27 10:15:31.101057017 +0000 UTC m=+0.184350074 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) Nov 27 05:15:31 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:15:31 localhost podman[331316]: 2025-11-27 10:15:31.153137986 +0000 UTC m=+0.229309311 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:15:31 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:15:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v650: 177 pgs: 177 active+clean; 212 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 157 KiB/s wr, 12 op/s Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.751 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.752 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.753 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.753 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:15:31 localhost nova_compute[284026]: 2025-11-27 10:15:31.869 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "4bbff30f-e3b8-4d51-834f-f8208de38bd5", "format": "json"}]: dispatch Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:32 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '4bbff30f-e3b8-4d51-834f-f8208de38bd5' of type subvolume Nov 27 05:15:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:32.561+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '4bbff30f-e3b8-4d51-834f-f8208de38bd5' of type subvolume Nov 27 05:15:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "4bbff30f-e3b8-4d51-834f-f8208de38bd5", "force": true, "format": "json"}]: dispatch Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/4bbff30f-e3b8-4d51-834f-f8208de38bd5'' moved to trashcan Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:4bbff30f-e3b8-4d51-834f-f8208de38bd5, vol_name:cephfs) < "" Nov 27 05:15:32 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:15:32 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:32 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:32 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:32 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e262 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:33 localhost nova_compute[284026]: 2025-11-27 10:15:33.159 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v651: 177 pgs: 177 active+clean; 212 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 157 KiB/s wr, 12 op/s Nov 27 05:15:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 e263: 6 total, 6 up, 6 in Nov 27 05:15:33 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:33 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:33 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:33 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:34 localhost nova_compute[284026]: 2025-11-27 10:15:34.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v653: 177 pgs: 177 active+clean; 212 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 1.1 KiB/s rd, 98 KiB/s wr, 9 op/s Nov 27 05:15:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "1bbe87b9-a95a-4636-a2b7-2abe81471266", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/1bbe87b9-a95a-4636-a2b7-2abe81471266/.meta.tmp' Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/1bbe87b9-a95a-4636-a2b7-2abe81471266/.meta.tmp' to config b'/volumes/_nogroup/1bbe87b9-a95a-4636-a2b7-2abe81471266/.meta' Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:35 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "1bbe87b9-a95a-4636-a2b7-2abe81471266", "format": "json"}]: dispatch Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:35 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:15:36 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:36 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:15:36 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:36 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:36 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:36 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:15:36 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:36 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:15:36 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.752 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.753 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.753 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.753 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.754 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:15:36 localhost nova_compute[284026]: 2025-11-27 10:15:36.872 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:37 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:15:37 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/1099397589' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.173 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.419s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.240 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.240 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:15:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v654: 177 pgs: 177 active+clean; 213 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 1005 B/s rd, 148 KiB/s wr, 12 op/s Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.470 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.471 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11047MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.472 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.472 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.803 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.803 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.804 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:15:37 localhost nova_compute[284026]: 2025-11-27 10:15:37.867 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:15:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.161 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:38 localhost ovn_controller[156436]: 2025-11-27T10:15:38Z|00350|memory_trim|INFO|Detected inactivity (last active 30002 ms ago): trimming memory Nov 27 05:15:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:15:38 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2683759798' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.354 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.488s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.361 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.377 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.380 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:15:38 localhost nova_compute[284026]: 2025-11-27 10:15:38.380 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.908s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:15:38 localhost podman[242678]: time="2025-11-27T10:15:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:15:38 localhost podman[242678]: @ - - [27/Nov/2025:10:15:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:15:38 localhost podman[242678]: @ - - [27/Nov/2025:10:15:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19754 "" "Go-http-client/1.1" Nov 27 05:15:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "1bbe87b9-a95a-4636-a2b7-2abe81471266", "format": "json"}]: dispatch Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:39.277+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '1bbe87b9-a95a-4636-a2b7-2abe81471266' of type subvolume Nov 27 05:15:39 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '1bbe87b9-a95a-4636-a2b7-2abe81471266' of type subvolume Nov 27 05:15:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "1bbe87b9-a95a-4636-a2b7-2abe81471266", "force": true, "format": "json"}]: dispatch Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/1bbe87b9-a95a-4636-a2b7-2abe81471266'' moved to trashcan Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:1bbe87b9-a95a-4636-a2b7-2abe81471266, vol_name:cephfs) < "" Nov 27 05:15:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v655: 177 pgs: 177 active+clean; 213 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 921 B/s rd, 136 KiB/s wr, 11 op/s Nov 27 05:15:39 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:39 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:39 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:39 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:39 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:40 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:40 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v656: 177 pgs: 177 active+clean; 213 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 126 KiB/s wr, 9 op/s Nov 27 05:15:41 localhost nova_compute[284026]: 2025-11-27 10:15:41.878 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:42 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [L] New memtable created with log file: #46. Immutable memtables: 3. Nov 27 05:15:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:42 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:42 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:15:42 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:42 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:42 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:43 localhost nova_compute[284026]: 2025-11-27 10:15:43.165 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v657: 177 pgs: 177 active+clean; 213 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 126 KiB/s wr, 9 op/s Nov 27 05:15:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:43.582 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:15:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:43.583 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:15:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:15:43.585 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:15:43 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:43 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:43 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:43 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:15:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:15:43 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:15:44 localhost systemd[1]: tmp-crun.syveQP.mount: Deactivated successfully. Nov 27 05:15:44 localhost podman[331409]: 2025-11-27 10:15:44.01305512 +0000 UTC m=+0.102980757 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) Nov 27 05:15:44 localhost podman[331409]: 2025-11-27 10:15:44.022001881 +0000 UTC m=+0.111927528 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, tcib_managed=true) Nov 27 05:15:44 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:15:44 localhost podman[331410]: 2025-11-27 10:15:44.10756997 +0000 UTC m=+0.193899340 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_id=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=multipathd) Nov 27 05:15:44 localhost podman[331410]: 2025-11-27 10:15:44.124038332 +0000 UTC m=+0.210367672 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, config_id=multipathd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) Nov 27 05:15:44 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.142 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.147 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4c2b7f1e-8e64-4ffc-bc64-1155eef9514b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.143273', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08bc0670-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'ecb0ed3cfa04191f5794ee3c5f516bb6921f1d88c20518ec622dfcbc23f3b833'}]}, 'timestamp': '2025-11-27 10:15:44.147769', '_unique_id': '52bd0ce234b74004bdf5684448221920'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.149 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.150 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.166 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '08cf724d-a715-4154-9aba-78150f964b3b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.150762', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08bef8c6-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': '9fb48e5fc8da61dbcdba130cc870c05270126fcc77f0af9c86c8d81101198cf4'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.150762', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08bf09ec-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': 'aadb712bb03c132c3397159fe4b9798d990e788639984bdede116dfb0f993c5f'}]}, 'timestamp': '2025-11-27 10:15:44.167374', '_unique_id': '45df9732b6ff465abd9792e526921a1a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.168 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.169 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.169 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '20d14178-7529-48fb-bc0b-638a1955a286', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.169882', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08bf7d00-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': '2eab34618659846f8a567e06af155eb53ebde586770d0bbbd0f3725ac8794548'}]}, 'timestamp': '2025-11-27 10:15:44.170347', '_unique_id': '31ab3d9cfe094234a65f71be953b1d8b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.172 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.172 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bf9806b2-6052-4e91-8af9-6575d5f96aed', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.172567', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08bfe61e-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'eceb3a81af240b9807bc45ea468732e3f0aad3522b9e4d5173fd380e771f84d4'}]}, 'timestamp': '2025-11-27 10:15:44.173038', '_unique_id': '50a7f5948a924e51bd31615536c24c78'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.173 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.175 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.175 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '92633a21-6a43-4d6a-bea4-23e4677180f5', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.175241', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08c04e4c-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': '59ea4d987f67f15f71364f345118c881632ca86aec6f3c57a1997ffe08bd02fd'}]}, 'timestamp': '2025-11-27 10:15:44.175738', '_unique_id': '8f23cf018be646b8bd692c7310d64ec1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.176 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.177 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.177 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '721f260c-c391-4e7a-8ae7-edc1a94b4586', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.177921', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08c0b6de-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'e03839d47ec863a643bde792cde4a86d377741c9137128ce0265392d3b99f85a'}]}, 'timestamp': '2025-11-27 10:15:44.178379', '_unique_id': '8580cbed25744e53a7bac38be70f0aba'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.179 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.180 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.180 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ece35cfa-c17f-4013-81d2-99d00cb94419', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.180602', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08c11ff2-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': '1f4b8cc296a2046125e8c6c84cbefef1be09a397da9aee4f059d71cd69c93372'}]}, 'timestamp': '2025-11-27 10:15:44.181069', '_unique_id': 'cc53bcc99e514f31ad8ea504a4b55ff1'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.182 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.183 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.183 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b89eff22-a41c-4a75-a5e7-b18c2d52ce32', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.183260', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08c18780-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'f76bba472ddb7c97c1f9aa895f15f21a6745c25da8cbfcfd272485e8c17a6e4f'}]}, 'timestamp': '2025-11-27 10:15:44.183753', '_unique_id': 'f9f138d89ada4bfa8f97477e489ec3b9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.184 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.185 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.211 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'e460c845-2cb2-4bfb-b95c-2c85a61be553', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.185913', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08c5ca34-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '6a6ec85f0c5ff966ffbb085b6f9c0db4958caefdd85549eb0a77df4a94adeda6'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.185913', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08c5dc72-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': 'cfed28a7e9fcc8ea25bf08e2cdeed67ff022a726b90af155fe2f757c36e608c0'}]}, 'timestamp': '2025-11-27 10:15:44.212084', '_unique_id': '9c69a320a5cc4895a5733467989d44d0'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.213 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.214 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.228 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 19260000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'aabdfd10-148b-46ac-b7ad-a8f57b460243', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 19260000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:15:44.214385', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '08c8754a-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.416607933, 'message_signature': 'f0dd2d2904b179fb93b138b66aec581ad86d9e8132eb58d46018124368185460'}]}, 'timestamp': '2025-11-27 10:15:44.229117', '_unique_id': '898d9df780894c34923ca1a1de96f93e'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.230 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.231 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.231 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.232 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.232 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9f311af3-174d-45e0-8f00-3c73f43e59cb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.232076', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08c8fc2c-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': '417a79ac0f8a46acc13018a86156d5fdc51fddd4a080dcb79ffc18438bf91d7e'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.232076', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08c90db6-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': 'd8099ef377ef6284b7af57ba2ccdfee5449bc27fb042b154cfeefa021709c4ac'}]}, 'timestamp': '2025-11-27 10:15:44.233001', '_unique_id': 'f15a0b1954dd4957a27b1ea2e6affee4'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.233 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.235 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.235 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.235 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '4d63fff4-66b1-42c7-9afe-c0d5d16cfb8c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.235193', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08c973e6-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '8e203a7e6296ffcef68770b89d0233bf158cb9a96808bbde4c6fa915b55253d4'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.235193', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08c98570-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '7fac2987e2323ccff3d07ab50a57f63f34dfdf874e86aa8c323805f9fdf2e252'}]}, 'timestamp': '2025-11-27 10:15:44.236067', '_unique_id': '968980f66fee4ec0a37447cf3bd19ade'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.238 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '521d9912-9c48-46a9-b2c6-4d133e4d36e8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.239065', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08ca0c52-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '73301853464d6f2051b1033a2b2e366a0dc91b18959cd2a70fe34c66c4e799b5'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.239065', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08ca1e04-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': 'ff87b8b241210deddb8e5535e41aa3ba76235ac56e6bc9fcb42dd6053d3e1b4d'}]}, 'timestamp': '2025-11-27 10:15:44.239972', '_unique_id': '7506a50daa594e6490c5026a63de1432'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.242 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '638174b9-523a-4b08-9036-a134daf33e8f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.242159', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08ca8420-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'f3b1c319bd67bf0314fcf821321ee8ffb21761647a107f967b7fbb6a335a6060'}]}, 'timestamp': '2025-11-27 10:15:44.242654', '_unique_id': 'b3c31b9c9e514611855aaa26ffa0228b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.244 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.244 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.245 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9ed272a9-d971-4367-8c59-4ec716297478', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.244931', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08caf0ae-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': 'd94b60fcac010ce5f061107c4debd44fa6b3d82d416af266834869ba31df94e2'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.244931', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08cb0224-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '9c07f222529bec2421c452241822b1a14ebe136db4630917a8db41ab08fbf6ea'}]}, 'timestamp': '2025-11-27 10:15:44.245819', '_unique_id': '0d134137d0b440b4b5ef485994d324cf'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.248 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '77f38cbb-1218-4563-b377-5513c023d865', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:15:44.247995', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '08cb693a-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.416607933, 'message_signature': '3029d50a3ec7e2c9952e0e75a8842bcc7489dcd208dc9911be7a43ad946db6cb'}]}, 'timestamp': '2025-11-27 10:15:44.248463', '_unique_id': '7ad12e5b29694b8ea190cd1e338d6ffd'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '53cc34ea-af8e-4045-b121-73ce96caeb0e', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.250594', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08cbcdc6-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': '35b2c3fe7d7fa5d04bf786eff2ab570e212e84da21c332466597edb7401920ab'}]}, 'timestamp': '2025-11-27 10:15:44.251051', '_unique_id': '50adc34675fb42299401954cbcf9c016'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.253 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '96d6ada8-6d08-43fa-8dea-04e6e24be6e0', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:15:44.253186', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '08cc32d4-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.331331743, 'message_signature': 'dcd1780056b0a7c9ceddb3f02a0a36f51aa247e10b672f4f71c8ad1614ac0ed7'}]}, 'timestamp': '2025-11-27 10:15:44.253672', '_unique_id': 'f9c393b30a944684883e14c1cf43eec9'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.255 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '09b6b48f-1231-4253-990a-c3fc2bec800f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.256028', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08cca2b4-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': '841c0ab4687a4d9a25402ae9e21f36427e74c90a117cd24294264b1f1d73db00'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.256028', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08ccb402-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.338817773, 'message_signature': '4b271e3abb1d8cc72013045f821577fb65377c8edfba9bb446d69d39b79543f2'}]}, 'timestamp': '2025-11-27 10:15:44.256919', '_unique_id': 'e88d62efb0d6443ea9afdd267363e22a'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.258 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'b6c156a2-d40c-4a25-bc72-1a4d7f2b7379', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.259215', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08cd1ea6-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '18deaf7eeb149aff9055b8014e2efaa79d6b1023823992dd5a8b5b29c31580a4'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.259215', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08cd3030-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '513a94e69ef1ffb06158e01fd7b95ba4ff5d9ec5d42d342600fdf9bb541e6a85'}]}, 'timestamp': '2025-11-27 10:15:44.260098', '_unique_id': 'c9cdb01ca02e4bdda1320cefa3e7bbef'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.262 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.262 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.262 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3cb54d5d-c052-4804-b677-7d4b6db774b1', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:15:44.262259', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '08cd9502-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': 'b02abbc6a95fe6cd9284ad11e4b337b77f62a04b374f478fbdd844875e69200c'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:15:44.262259', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '08cda682-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12774.373969638, 'message_signature': '59f92d16baaabcac87380730646832fbdc0e44a7091e668738bcb40438a2f748'}]}, 'timestamp': '2025-11-27 10:15:44.263126', '_unique_id': '08d920243bfa4dd69a560a240a82c98b'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:15:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:15:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.382 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.383 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.383 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.462 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.462 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.463 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:15:44 localhost nova_compute[284026]: 2025-11-27 10:15:44.463 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:15:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:15:44 Nov 27 05:15:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:15:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:15:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['volumes', 'manila_metadata', 'images', '.mgr', 'manila_data', 'backups', 'vms'] Nov 27 05:15:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:15:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 1.0905220547180346e-06 of space, bias 1.0, pg target 0.00021701388888888888 quantized to 32 (current 32) Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:15:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0015223687883863762 of space, bias 4.0, pg target 1.2118055555555556 quantized to 16 (current 16) Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:15:44 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:15:45 localhost nova_compute[284026]: 2025-11-27 10:15:45.075 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:15:45 localhost nova_compute[284026]: 2025-11-27 10:15:45.092 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:15:45 localhost nova_compute[284026]: 2025-11-27 10:15:45.093 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:15:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v658: 177 pgs: 177 active+clean; 213 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 512 B/s rd, 105 KiB/s wr, 8 op/s Nov 27 05:15:46 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:15:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:46 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:46 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:46 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:46 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:46 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:46 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:46 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:46 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:46 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:46 localhost nova_compute[284026]: 2025-11-27 10:15:46.880 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v659: 177 pgs: 177 active+clean; 214 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 135 KiB/s wr, 9 op/s Nov 27 05:15:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:48 localhost nova_compute[284026]: 2025-11-27 10:15:48.168 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v660: 177 pgs: 177 active+clean; 214 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 88 KiB/s wr, 6 op/s Nov 27 05:15:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:15:49 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:49 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:15:49 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:49 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:49 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:49 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:15:49 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:49 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:15:49 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:15:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "05c4ecf5-7dfa-487d-ae5f-6a485198f7dd", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/05c4ecf5-7dfa-487d-ae5f-6a485198f7dd/.meta.tmp' Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/05c4ecf5-7dfa-487d-ae5f-6a485198f7dd/.meta.tmp' to config b'/volumes/_nogroup/05c4ecf5-7dfa-487d-ae5f-6a485198f7dd/.meta' Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v661: 177 pgs: 177 active+clean; 214 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 128 KiB/s wr, 10 op/s Nov 27 05:15:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "05c4ecf5-7dfa-487d-ae5f-6a485198f7dd", "format": "json"}]: dispatch Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta' Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "format": "json"}]: dispatch Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:51 localhost nova_compute[284026]: 2025-11-27 10:15:51.882 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:15:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:15:51 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:15:52 localhost podman[331448]: 2025-11-27 10:15:52.003631157 +0000 UTC m=+0.087284217 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, com.redhat.component=ubi9-minimal-container, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.buildah.version=1.33.7, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, distribution-scope=public, architecture=x86_64, container_name=openstack_network_exporter, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., io.openshift.tags=minimal rhel9, version=9.6, build-date=2025-08-20T13:12:41, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., managed_by=edpm_ansible, io.openshift.expose-services=, maintainer=Red Hat, Inc., name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, config_id=edpm) Nov 27 05:15:52 localhost podman[331448]: 2025-11-27 10:15:52.01681922 +0000 UTC m=+0.100472320 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, architecture=x86_64, vcs-type=git, vendor=Red Hat, Inc., config_id=edpm, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.expose-services=, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, maintainer=Red Hat, Inc., version=9.6, io.buildah.version=1.33.7, release=1755695350, distribution-scope=public, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, build-date=2025-08-20T13:12:41, name=ubi9-minimal) Nov 27 05:15:52 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:15:52 localhost podman[331447]: 2025-11-27 10:15:52.099478461 +0000 UTC m=+0.184118957 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:15:52 localhost podman[331447]: 2025-11-27 10:15:52.110846947 +0000 UTC m=+0.195487443 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:15:52 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:15:52 localhost podman[331446]: 2025-11-27 10:15:52.159501053 +0000 UTC m=+0.249645207 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:15:52 localhost podman[331446]: 2025-11-27 10:15:52.194849393 +0000 UTC m=+0.284993547 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm) Nov 27 05:15:52 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:15:52 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:15:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:52 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:52 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:52 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:52 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:52 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:53 localhost nova_compute[284026]: 2025-11-27 10:15:53.171 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta' Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "format": "json"}]: dispatch Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v662: 177 pgs: 177 active+clean; 214 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 71 KiB/s wr, 6 op/s Nov 27 05:15:53 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:53 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:15:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "snap_name": "736c2489-7825-4818-9921-16f731e22c5d", "format": "json"}]: dispatch Nov 27 05:15:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:736c2489-7825-4818-9921-16f731e22c5d, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:736c2489-7825-4818-9921-16f731e22c5d, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:15:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v663: 177 pgs: 177 active+clean; 214 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 71 KiB/s wr, 6 op/s Nov 27 05:15:55 localhost openstack_network_exporter[244641]: ERROR 10:15:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:15:55 localhost openstack_network_exporter[244641]: ERROR 10:15:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:15:55 localhost openstack_network_exporter[244641]: ERROR 10:15:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:15:55 localhost openstack_network_exporter[244641]: ERROR 10:15:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:15:55 localhost openstack_network_exporter[244641]: Nov 27 05:15:55 localhost openstack_network_exporter[244641]: ERROR 10:15:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:15:55 localhost openstack_network_exporter[244641]: Nov 27 05:15:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:56 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:15:56 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:15:56 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "snap_name": "b0cb7398-c338-466a-859e-ce1ca7407333", "format": "json"}]: dispatch Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:56 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:15:56 localhost nova_compute[284026]: 2025-11-27 10:15:56.883 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:56 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:56 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:56 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:15:56 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:15:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v664: 177 pgs: 177 active+clean; 215 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 170 B/s rd, 123 KiB/s wr, 9 op/s Nov 27 05:15:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e263 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:15:58 localhost nova_compute[284026]: 2025-11-27 10:15:58.173 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:15:58 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "4742f96a-82d3-41e9-9637-edb4d63125ad", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/4742f96a-82d3-41e9-9637-edb4d63125ad/.meta.tmp' Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/4742f96a-82d3-41e9-9637-edb4d63125ad/.meta.tmp' to config b'/volumes/_nogroup/4742f96a-82d3-41e9-9637-edb4d63125ad/.meta' Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:15:58 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "4742f96a-82d3-41e9-9637-edb4d63125ad", "format": "json"}]: dispatch Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:15:58 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "05c4ecf5-7dfa-487d-ae5f-6a485198f7dd", "format": "json"}]: dispatch Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:15:59.411+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '05c4ecf5-7dfa-487d-ae5f-6a485198f7dd' of type subvolume Nov 27 05:15:59 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '05c4ecf5-7dfa-487d-ae5f-6a485198f7dd' of type subvolume Nov 27 05:15:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "05c4ecf5-7dfa-487d-ae5f-6a485198f7dd", "force": true, "format": "json"}]: dispatch Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/05c4ecf5-7dfa-487d-ae5f-6a485198f7dd'' moved to trashcan Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:05c4ecf5-7dfa-487d-ae5f-6a485198f7dd, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v665: 177 pgs: 177 active+clean; 215 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 170 B/s rd, 92 KiB/s wr, 7 op/s Nov 27 05:15:59 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:15:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:59 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:15:59 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:15:59 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:59 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:15:59 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:15:59 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:59 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:15:59 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:00 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "snap_name": "b0cb7398-c338-466a-859e-ce1ca7407333_254077d7-0fb4-4a9e-8456-c30c6dd8823a", "force": true, "format": "json"}]: dispatch Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333_254077d7-0fb4-4a9e-8456-c30c6dd8823a, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta' Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333_254077d7-0fb4-4a9e-8456-c30c6dd8823a, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:00 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "snap_name": "b0cb7398-c338-466a-859e-ce1ca7407333", "force": true, "format": "json"}]: dispatch Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta.tmp' to config b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41/.meta' Nov 27 05:16:00 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:b0cb7398-c338-466a-859e-ce1ca7407333, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v666: 177 pgs: 177 active+clean; 216 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 167 KiB/s wr, 12 op/s Nov 27 05:16:01 localhost nova_compute[284026]: 2025-11-27 10:16:01.918 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:16:01 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:16:02 localhost systemd[1]: tmp-crun.PKOliN.mount: Deactivated successfully. Nov 27 05:16:02 localhost podman[331506]: 2025-11-27 10:16:02.043062325 +0000 UTC m=+0.099352249 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:16:02 localhost podman[331506]: 2025-11-27 10:16:02.055950552 +0000 UTC m=+0.112240466 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:16:02 localhost podman[331505]: 2025-11-27 10:16:02.013581264 +0000 UTC m=+0.076080184 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_managed=true, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_controller, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:16:02 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:16:02 localhost podman[331505]: 2025-11-27 10:16:02.099854712 +0000 UTC m=+0.162353612 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.schema-version=1.0, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=ovn_controller) Nov 27 05:16:02 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:16:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "4742f96a-82d3-41e9-9637-edb4d63125ad", "format": "json"}]: dispatch Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:4742f96a-82d3-41e9-9637-edb4d63125ad, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:4742f96a-82d3-41e9-9637-edb4d63125ad, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:02 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:02.204+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '4742f96a-82d3-41e9-9637-edb4d63125ad' of type subvolume Nov 27 05:16:02 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '4742f96a-82d3-41e9-9637-edb4d63125ad' of type subvolume Nov 27 05:16:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "4742f96a-82d3-41e9-9637-edb4d63125ad", "force": true, "format": "json"}]: dispatch Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/4742f96a-82d3-41e9-9637-edb4d63125ad'' moved to trashcan Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:4742f96a-82d3-41e9-9637-edb4d63125ad, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e264 e264: 6 total, 6 up, 6 in Nov 27 05:16:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e264 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:03 localhost nova_compute[284026]: 2025-11-27 10:16:03.206 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice", "format": "json"} v 0) Nov 27 05:16:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:16:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice"} v 0) Nov 27 05:16:03 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice", "format": "json"}]: dispatch Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v668: 177 pgs: 177 active+clean; 216 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 153 KiB/s wr, 11 op/s Nov 27 05:16:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "format": "json"}]: dispatch Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:03.615+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'fb2c13ad-9351-41be-ac91-5e74d14ded41' of type subvolume Nov 27 05:16:03 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'fb2c13ad-9351-41be-ac91-5e74d14ded41' of type subvolume Nov 27 05:16:03 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "fb2c13ad-9351-41be-ac91-5e74d14ded41", "force": true, "format": "json"}]: dispatch Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/fb2c13ad-9351-41be-ac91-5e74d14ded41'' moved to trashcan Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:03 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:fb2c13ad-9351-41be-ac91-5e74d14ded41, vol_name:cephfs) < "" Nov 27 05:16:04 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice", "format": "json"} : dispatch Nov 27 05:16:04 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:16:04 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice"} : dispatch Nov 27 05:16:04 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice"}]': finished Nov 27 05:16:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v669: 177 pgs: 177 active+clean; 216 MiB data, 1.2 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 205 KiB/s wr, 13 op/s Nov 27 05:16:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "5b78a3da-4cc0-41af-9868-46cefb3d016e", "size": 2147483648, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/5b78a3da-4cc0-41af-9868-46cefb3d016e/.meta.tmp' Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/5b78a3da-4cc0-41af-9868-46cefb3d016e/.meta.tmp' to config b'/volumes/_nogroup/5b78a3da-4cc0-41af-9868-46cefb3d016e/.meta' Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:2147483648, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "5b78a3da-4cc0-41af-9868-46cefb3d016e", "format": "json"}]: dispatch Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:06 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:16:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:06 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:16:06 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:06 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:06 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:06 localhost nova_compute[284026]: 2025-11-27 10:16:06.920 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:07 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:07 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v670: 177 pgs: 177 active+clean; 216 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 716 B/s rd, 146 KiB/s wr, 11 op/s Nov 27 05:16:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e264 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:08 localhost nova_compute[284026]: 2025-11-27 10:16:08.208 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 e265: 6 total, 6 up, 6 in Nov 27 05:16:08 localhost podman[242678]: time="2025-11-27T10:16:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:16:08 localhost podman[242678]: @ - - [27/Nov/2025:10:16:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:16:08 localhost podman[242678]: @ - - [27/Nov/2025:10:16:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19753 "" "Go-http-client/1.1" Nov 27 05:16:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "5b78a3da-4cc0-41af-9868-46cefb3d016e", "format": "json"}]: dispatch Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:09 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:09.416+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5b78a3da-4cc0-41af-9868-46cefb3d016e' of type subvolume Nov 27 05:16:09 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '5b78a3da-4cc0-41af-9868-46cefb3d016e' of type subvolume Nov 27 05:16:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "5b78a3da-4cc0-41af-9868-46cefb3d016e", "force": true, "format": "json"}]: dispatch Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/5b78a3da-4cc0-41af-9868-46cefb3d016e'' moved to trashcan Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:5b78a3da-4cc0-41af-9868-46cefb3d016e, vol_name:cephfs) < "" Nov 27 05:16:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v672: 177 pgs: 177 active+clean; 216 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 767 B/s rd, 70 KiB/s wr, 5 op/s Nov 27 05:16:09 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:16:09 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:16:10 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:10 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:16:10 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:16:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:11 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:11 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:11 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:11 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:16:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v673: 177 pgs: 177 active+clean; 217 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 974 B/s rd, 169 KiB/s wr, 12 op/s Nov 27 05:16:11 localhost nova_compute[284026]: 2025-11-27 10:16:11.921 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "a5b79619-245a-42de-9e26-e5307a5c816e", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a5b79619-245a-42de-9e26-e5307a5c816e/.meta.tmp' Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a5b79619-245a-42de-9e26-e5307a5c816e/.meta.tmp' to config b'/volumes/_nogroup/a5b79619-245a-42de-9e26-e5307a5c816e/.meta' Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "a5b79619-245a-42de-9e26-e5307a5c816e", "format": "json"}]: dispatch Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:13 localhost nova_compute[284026]: 2025-11-27 10:16:13.210 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:13 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:16:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:13 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice_bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:16:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:13.430 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=26, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=25) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:16:13 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:13.431 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 8 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:16:13 localhost nova_compute[284026]: 2025-11-27 10:16:13.430 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:13 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v674: 177 pgs: 177 active+clean; 217 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 818 B/s rd, 142 KiB/s wr, 10 op/s Nov 27 05:16:13 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice_bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:16:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:14 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:14 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice_bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:14 localhost podman[331591]: 2025-11-27 10:16:14.189780944 +0000 UTC m=+0.075053348 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_id=ovn_metadata_agent, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0) Nov 27 05:16:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:16:14 localhost podman[331591]: 2025-11-27 10:16:14.23505162 +0000 UTC m=+0.120324034 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent) Nov 27 05:16:14 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:16:14 localhost podman[331609]: 2025-11-27 10:16:14.302244475 +0000 UTC m=+0.087828920 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:16:14 localhost podman[331609]: 2025-11-27 10:16:14.338938191 +0000 UTC m=+0.124522646 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, container_name=multipathd, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, io.buildah.version=1.41.3) Nov 27 05:16:14 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:16:14 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:16:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:16:14 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:16:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:16:14 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 079659a5-fbbe-486d-9a6c-ddb1a1a078af (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:16:14 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 079659a5-fbbe-486d-9a6c-ddb1a1a078af (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:16:14 localhost ceph-mgr[290377]: [progress INFO root] Completed event 079659a5-fbbe-486d-9a6c-ddb1a1a078af (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:16:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:16:14 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:16:14 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:16:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:16:15 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:16:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:16:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:16:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v675: 177 pgs: 177 active+clean; 217 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 130 KiB/s wr, 10 op/s Nov 27 05:16:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a5b79619-245a-42de-9e26-e5307a5c816e", "format": "json"}]: dispatch Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a5b79619-245a-42de-9e26-e5307a5c816e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a5b79619-245a-42de-9e26-e5307a5c816e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:16.385+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a5b79619-245a-42de-9e26-e5307a5c816e' of type subvolume Nov 27 05:16:16 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a5b79619-245a-42de-9e26-e5307a5c816e' of type subvolume Nov 27 05:16:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "a5b79619-245a-42de-9e26-e5307a5c816e", "force": true, "format": "json"}]: dispatch Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/a5b79619-245a-42de-9e26-e5307a5c816e'' moved to trashcan Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a5b79619-245a-42de-9e26-e5307a5c816e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} v 0) Nov 27 05:16:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice_bob"} v 0) Nov 27 05:16:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice_bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice_bob", "format": "json"}]: dispatch Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice_bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:16 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice_bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:16 localhost nova_compute[284026]: 2025-11-27 10:16:16.923 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice_bob", "format": "json"} : dispatch Nov 27 05:16:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice_bob"} : dispatch Nov 27 05:16:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice_bob"}]': finished Nov 27 05:16:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v676: 177 pgs: 177 active+clean; 217 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 128 KiB/s wr, 9 op/s Nov 27 05:16:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:18 localhost nova_compute[284026]: 2025-11-27 10:16:18.214 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v677: 177 pgs: 177 active+clean; 217 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 466 B/s rd, 116 KiB/s wr, 8 op/s Nov 27 05:16:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "a146106c-3837-4a7a-bb06-9a6a9fde2428", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/a146106c-3837-4a7a-bb06-9a6a9fde2428/.meta.tmp' Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/a146106c-3837-4a7a-bb06-9a6a9fde2428/.meta.tmp' to config b'/volumes/_nogroup/a146106c-3837-4a7a-bb06-9a6a9fde2428/.meta' Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "a146106c-3837-4a7a-bb06-9a6a9fde2428", "format": "json"}]: dispatch Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:19 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:16:19 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:19 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:16:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:20 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:20 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:20 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:21 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:21.434 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '26'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:16:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v678: 177 pgs: 177 active+clean; 218 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 682 B/s rd, 176 KiB/s wr, 13 op/s Nov 27 05:16:21 localhost nova_compute[284026]: 2025-11-27 10:16:21.925 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:16:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:16:22 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:16:22 localhost podman[331680]: 2025-11-27 10:16:22.998916267 +0000 UTC m=+0.088420386 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:16:23 localhost podman[331679]: 2025-11-27 10:16:23.048675514 +0000 UTC m=+0.139949830 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, tcib_managed=true, container_name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, managed_by=edpm_ansible) Nov 27 05:16:23 localhost podman[331679]: 2025-11-27 10:16:23.063017789 +0000 UTC m=+0.154292095 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ceilometer_agent_compute, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, config_id=edpm, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:16:23 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:16:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:23 localhost podman[331681]: 2025-11-27 10:16:23.158641778 +0000 UTC m=+0.244583101 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, build-date=2025-08-20T13:12:41, architecture=x86_64, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, url=https://catalog.redhat.com/en/search?searchType=containers, name=ubi9-minimal, config_id=edpm, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=Red Hat, Inc., release=1755695350, vendor=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., managed_by=edpm_ansible, distribution-scope=public, vcs-type=git) Nov 27 05:16:23 localhost podman[331680]: 2025-11-27 10:16:23.166945171 +0000 UTC m=+0.256449250 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:16:23 localhost podman[331681]: 2025-11-27 10:16:23.17807999 +0000 UTC m=+0.264021323 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., maintainer=Red Hat, Inc., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, vendor=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.expose-services=, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.openshift.tags=minimal rhel9, managed_by=edpm_ansible, version=9.6, io.buildah.version=1.33.7, container_name=openstack_network_exporter, build-date=2025-08-20T13:12:41, config_id=edpm, architecture=x86_64, name=ubi9-minimal, release=1755695350, distribution-scope=public, vcs-type=git, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., url=https://catalog.redhat.com/en/search?searchType=containers, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:16:23 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:16:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "a146106c-3837-4a7a-bb06-9a6a9fde2428", "format": "json"}]: dispatch Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:23 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a146106c-3837-4a7a-bb06-9a6a9fde2428' of type subvolume Nov 27 05:16:23 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:23.188+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'a146106c-3837-4a7a-bb06-9a6a9fde2428' of type subvolume Nov 27 05:16:23 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:16:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "a146106c-3837-4a7a-bb06-9a6a9fde2428", "force": true, "format": "json"}]: dispatch Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/a146106c-3837-4a7a-bb06-9a6a9fde2428'' moved to trashcan Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:a146106c-3837-4a7a-bb06-9a6a9fde2428, vol_name:cephfs) < "" Nov 27 05:16:23 localhost nova_compute[284026]: 2025-11-27 10:16:23.216 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:23 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v679: 177 pgs: 177 active+clean; 218 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 104 KiB/s wr, 8 op/s Nov 27 05:16:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:16:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:16:23 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:23 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:23 localhost systemd[1]: tmp-crun.5Rbd0F.mount: Deactivated successfully. Nov 27 05:16:24 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:16:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:24 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:24 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:24 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:24 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:24 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:24 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:16:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v680: 177 pgs: 177 active+clean; 219 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 135 KiB/s wr, 9 op/s Nov 27 05:16:25 localhost openstack_network_exporter[244641]: ERROR 10:16:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:16:25 localhost openstack_network_exporter[244641]: ERROR 10:16:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:16:25 localhost openstack_network_exporter[244641]: ERROR 10:16:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:16:25 localhost openstack_network_exporter[244641]: ERROR 10:16:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:16:25 localhost openstack_network_exporter[244641]: Nov 27 05:16:25 localhost openstack_network_exporter[244641]: ERROR 10:16:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:16:25 localhost openstack_network_exporter[244641]: Nov 27 05:16:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "9305d31e-0e33-43d3-afb3-8986a8f125be", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/9305d31e-0e33-43d3-afb3-8986a8f125be/.meta.tmp' Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/9305d31e-0e33-43d3-afb3-8986a8f125be/.meta.tmp' to config b'/volumes/_nogroup/9305d31e-0e33-43d3-afb3-8986a8f125be/.meta' Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "9305d31e-0e33-43d3-afb3-8986a8f125be", "format": "json"}]: dispatch Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "r", "format": "json"}]: dispatch Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:16:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:26 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID alice bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:16:26 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:26 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:r, auth_id:alice bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:26 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:26 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:26 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:26 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.alice bob", "caps": ["mds", "allow r path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow r pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:26 localhost nova_compute[284026]: 2025-11-27 10:16:26.928 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:16:27 localhost sshd[331737]: main: sshd: ssh-rsa algorithm is disabled Nov 27 05:16:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/.meta.tmp' Nov 27 05:16:27 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/.meta.tmp' to config b'/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/.meta' Nov 27 05:16:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:16:27 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "format": "json"}]: dispatch Nov 27 05:16:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:16:27 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:16:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v681: 177 pgs: 177 active+clean; 219 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 102 KiB/s wr, 7 op/s Nov 27 05:16:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:28 localhost nova_compute[284026]: 2025-11-27 10:16:28.219 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v682: 177 pgs: 177 active+clean; 219 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 101 KiB/s wr, 7 op/s Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.alice bob", "format": "json"} v 0) Nov 27 05:16:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:30 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.alice bob"} v 0) Nov 27 05:16:30 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:alice bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "alice bob", "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=alice bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:alice bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:30 localhost nova_compute[284026]: 2025-11-27 10:16:30.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/.meta.tmp' Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/.meta.tmp' to config b'/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/.meta' Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "9305d31e-0e33-43d3-afb3-8986a8f125be", "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:9305d31e-0e33-43d3-afb3-8986a8f125be, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:9305d31e-0e33-43d3-afb3-8986a8f125be, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:30.780+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '9305d31e-0e33-43d3-afb3-8986a8f125be' of type subvolume Nov 27 05:16:30 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '9305d31e-0e33-43d3-afb3-8986a8f125be' of type subvolume Nov 27 05:16:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "9305d31e-0e33-43d3-afb3-8986a8f125be", "force": true, "format": "json"}]: dispatch Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/9305d31e-0e33-43d3-afb3-8986a8f125be'' moved to trashcan Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:9305d31e-0e33-43d3-afb3-8986a8f125be, vol_name:cephfs) < "" Nov 27 05:16:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.alice bob", "format": "json"} : dispatch Nov 27 05:16:30 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.alice bob"} : dispatch Nov 27 05:16:30 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.alice bob"}]': finished Nov 27 05:16:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v683: 177 pgs: 177 active+clean; 219 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 161 KiB/s wr, 12 op/s Nov 27 05:16:31 localhost nova_compute[284026]: 2025-11-27 10:16:31.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:31 localhost nova_compute[284026]: 2025-11-27 10:16:31.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:31 localhost nova_compute[284026]: 2025-11-27 10:16:31.931 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:16:32 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:16:32 localhost podman[331741]: 2025-11-27 10:16:32.988004135 +0000 UTC m=+0.082095767 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:16:33 localhost podman[331741]: 2025-11-27 10:16:32.999952095 +0000 UTC m=+0.094043697 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:16:33 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:16:33 localhost podman[331740]: 2025-11-27 10:16:33.088110603 +0000 UTC m=+0.184988910 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, config_id=ovn_controller, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:16:33 localhost podman[331740]: 2025-11-27 10:16:33.124391868 +0000 UTC m=+0.221270135 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_id=ovn_controller, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0) Nov 27 05:16:33 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:16:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e265 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:33 localhost nova_compute[284026]: 2025-11-27 10:16:33.221 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v684: 177 pgs: 177 active+clean; 219 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 91 KiB/s wr, 7 op/s Nov 27 05:16:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.bob", "format": "json"} v 0) Nov 27 05:16:33 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:33 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID bob with tenant c9c9a216ebe0444884e919df70e22b55 Nov 27 05:16:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:33 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:bob, format:json, prefix:fs subvolume authorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:33 localhost nova_compute[284026]: 2025-11-27 10:16:33.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:33 localhost nova_compute[284026]: 2025-11-27 10:16:33.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:33 localhost nova_compute[284026]: 2025-11-27 10:16:33.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:16:33 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "auth_id": "Joe", "tenant_id": "e5da946dcab84022ab5dfc2c601c61e9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:33 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:Joe, format:json, prefix:fs subvolume authorize, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:16:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.Joe", "format": "json"} v 0) Nov 27 05:16:34 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:34 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID Joe with tenant e5da946dcab84022ab5dfc2c601c61e9 Nov 27 05:16:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:34 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:34 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:34 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.Joe", "caps": ["mds", "allow rw path=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982", "osd", "allow rw pool=manila_data namespace=fsvolumens_15efeb80-6730-4067-836c-71b94433ea7a", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:34 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.Joe", "caps": ["mds", "allow rw path=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982", "osd", "allow rw pool=manila_data namespace=fsvolumens_15efeb80-6730-4067-836c-71b94433ea7a", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:Joe, format:json, prefix:fs subvolume authorize, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:16:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "snap_name": "736c2489-7825-4818-9921-16f731e22c5d_1d4a62d4-204c-47af-a083-d067f31893c4", "force": true, "format": "json"}]: dispatch Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:736c2489-7825-4818-9921-16f731e22c5d_1d4a62d4-204c-47af-a083-d067f31893c4, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta' Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:736c2489-7825-4818-9921-16f731e22c5d_1d4a62d4-204c-47af-a083-d067f31893c4, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "snap_name": "736c2489-7825-4818-9921-16f731e22c5d", "force": true, "format": "json"}]: dispatch Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:736c2489-7825-4818-9921-16f731e22c5d, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta.tmp' to config b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42/.meta' Nov 27 05:16:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:736c2489-7825-4818-9921-16f731e22c5d, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:34 localhost nova_compute[284026]: 2025-11-27 10:16:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:35 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:35 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.Joe", "caps": ["mds", "allow rw path=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982", "osd", "allow rw pool=manila_data namespace=fsvolumens_15efeb80-6730-4067-836c-71b94433ea7a", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:35 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.Joe", "caps": ["mds", "allow rw path=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982", "osd", "allow rw pool=manila_data namespace=fsvolumens_15efeb80-6730-4067-836c-71b94433ea7a", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:35 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.Joe", "caps": ["mds", "allow rw path=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982", "osd", "allow rw pool=manila_data namespace=fsvolumens_15efeb80-6730-4067-836c-71b94433ea7a", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v685: 177 pgs: 177 active+clean; 220 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 171 KiB/s wr, 11 op/s Nov 27 05:16:36 localhost nova_compute[284026]: 2025-11-27 10:16:36.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:36 localhost nova_compute[284026]: 2025-11-27 10:16:36.933 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/.meta.tmp' Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/.meta.tmp' to config b'/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/.meta' Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v686: 177 pgs: 177 active+clean; 220 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 140 KiB/s wr, 11 op/s Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b494d69a-bce7-42b4-8699-8b8d8bf39e42' of type subvolume Nov 27 05:16:37 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:37.591+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b494d69a-bce7-42b4-8699-8b8d8bf39e42' of type subvolume Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "b494d69a-bce7-42b4-8699-8b8d8bf39e42", "force": true, "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/b494d69a-bce7-42b4-8699-8b8d8bf39e42'' moved to trashcan Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b494d69a-bce7-42b4-8699-8b8d8bf39e42, vol_name:cephfs) < "" Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.750 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.751 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.751 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:16:37 localhost nova_compute[284026]: 2025-11-27 10:16:37.752 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/.meta.tmp' Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/.meta.tmp' to config b'/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/.meta' Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "format": "json"}]: dispatch Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:37 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e266 e266: 6 total, 6 up, 6 in Nov 27 05:16:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e266 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.225 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:16:38 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2668112415' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.255 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.503s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.323 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.323 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.554 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.556 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11040MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.556 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.557 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.623 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.623 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.624 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:16:38 localhost nova_compute[284026]: 2025-11-27 10:16:38.666 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:16:38 localhost podman[242678]: time="2025-11-27T10:16:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:16:38 localhost podman[242678]: @ - - [27/Nov/2025:10:16:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:16:38 localhost podman[242678]: @ - - [27/Nov/2025:10:16:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19750 "" "Go-http-client/1.1" Nov 27 05:16:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:16:39 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/853123308' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:16:39 localhost nova_compute[284026]: 2025-11-27 10:16:39.195 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.529s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:16:39 localhost nova_compute[284026]: 2025-11-27 10:16:39.202 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:16:39 localhost nova_compute[284026]: 2025-11-27 10:16:39.221 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:16:39 localhost nova_compute[284026]: 2025-11-27 10:16:39.224 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:16:39 localhost nova_compute[284026]: 2025-11-27 10:16:39.224 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.667s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:16:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v688: 177 pgs: 177 active+clean; 220 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 167 KiB/s wr, 12 op/s Nov 27 05:16:40 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "Joe", "tenant_id": "31016ff904ec4126b31d88584782c763", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:Joe, format:json, prefix:fs subvolume authorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:16:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.Joe", "format": "json"} v 0) Nov 27 05:16:40 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:40 localhost ceph-mgr[290377]: [volumes ERROR volumes.fs.operations.versions.subvolume_v1] auth ID: Joe is already in use Nov 27 05:16:40 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:Joe, format:json, prefix:fs subvolume authorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:16:40 localhost ceph-mgr[290377]: mgr.server reply reply (1) Operation not permitted auth ID: Joe is already in use Nov 27 05:16:40 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:40.318+0000 7f54df954640 -1 mgr.server reply reply (1) Operation not permitted auth ID: Joe is already in use Nov 27 05:16:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "auth_id": "bob", "tenant_id": "c9c9a216ebe0444884e919df70e22b55", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:bob, format:json, prefix:fs subvolume authorize, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.bob", "format": "json"} v 0) Nov 27 05:16:41 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth caps", "entity": "client.bob", "caps": ["mon", "allow r", "mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177,allow rw path=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e,allow rw pool=manila_data namespace=fsvolumens_7aeed949-6381-43a8-bac5-b37209c017f2"]} v 0) Nov 27 05:16:41 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mon", "allow r", "mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177,allow rw path=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e,allow rw pool=manila_data namespace=fsvolumens_7aeed949-6381-43a8-bac5-b37209c017f2"]} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.bob", "format": "json"} v 0) Nov 27 05:16:41 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:bob, format:json, prefix:fs subvolume authorize, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, tenant_id:c9c9a216ebe0444884e919df70e22b55, vol_name:cephfs) < "" Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mon", "allow r", "mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177,allow rw path=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e,allow rw pool=manila_data namespace=fsvolumens_7aeed949-6381-43a8-bac5-b37209c017f2"]} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mon", "allow r", "mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177,allow rw path=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e,allow rw pool=manila_data namespace=fsvolumens_7aeed949-6381-43a8-bac5-b37209c017f2"]} : dispatch Nov 27 05:16:41 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth caps", "entity": "client.bob", "caps": ["mon", "allow r", "mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177,allow rw path=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e,allow rw pool=manila_data namespace=fsvolumens_7aeed949-6381-43a8-bac5-b37209c017f2"]}]': finished Nov 27 05:16:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v689: 177 pgs: 177 active+clean; 221 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 614 B/s rd, 150 KiB/s wr, 9 op/s Nov 27 05:16:41 localhost nova_compute[284026]: 2025-11-27 10:16:41.936 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e266 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:43 localhost nova_compute[284026]: 2025-11-27 10:16:43.228 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v690: 177 pgs: 177 active+clean; 221 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 614 B/s rd, 150 KiB/s wr, 9 op/s Nov 27 05:16:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:43.583 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:16:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:43.584 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:16:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:16:43.584 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:16:43 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "tempest-cephx-id-471164033", "tenant_id": "31016ff904ec4126b31d88584782c763", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume authorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:16:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} v 0) Nov 27 05:16:43 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} : dispatch Nov 27 05:16:43 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID tempest-cephx-id-471164033 with tenant 31016ff904ec4126b31d88584782c763 Nov 27 05:16:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-471164033", "caps": ["mds", "allow rw path=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af", "osd", "allow rw pool=manila_data namespace=fsvolumens_542efcb9-ee89-45b6-928b-9c3f42f649e6", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:16:43 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-471164033", "caps": ["mds", "allow rw path=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af", "osd", "allow rw pool=manila_data namespace=fsvolumens_542efcb9-ee89-45b6-928b-9c3f42f649e6", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:43 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume authorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:16:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e267 e267: 6 total, 6 up, 6 in Nov 27 05:16:44 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} : dispatch Nov 27 05:16:44 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-471164033", "caps": ["mds", "allow rw path=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af", "osd", "allow rw pool=manila_data namespace=fsvolumens_542efcb9-ee89-45b6-928b-9c3f42f649e6", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:44 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-471164033", "caps": ["mds", "allow rw path=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af", "osd", "allow rw pool=manila_data namespace=fsvolumens_542efcb9-ee89-45b6-928b-9c3f42f649e6", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:16:44 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.tempest-cephx-id-471164033", "caps": ["mds", "allow rw path=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af", "osd", "allow rw pool=manila_data namespace=fsvolumens_542efcb9-ee89-45b6-928b-9c3f42f649e6", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:16:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "auth_id": "bob", "format": "json"}]: dispatch Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:bob, format:json, prefix:fs subvolume deauthorize, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:16:44 Nov 27 05:16:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:16:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:16:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['.mgr', 'vms', 'manila_data', 'manila_metadata', 'images', 'backups', 'volumes'] Nov 27 05:16:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:16:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.bob", "format": "json"} v 0) Nov 27 05:16:44 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:44 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth caps", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e"]} v 0) Nov 27 05:16:44 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e"]} : dispatch Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:bob, format:json, prefix:fs subvolume deauthorize, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:44 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "auth_id": "bob", "format": "json"}]: dispatch Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:bob, format:json, prefix:fs subvolume evict, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=bob, client_metadata.root=/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2/06736ea0-c3bb-4180-a111-ed8011cfb590 Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:bob, format:json, prefix:fs subvolume evict, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:16:44 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:16:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:16:44 localhost podman[331832]: 2025-11-27 10:16:44.740481031 +0000 UTC m=+0.097116480 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:16:44 localhost podman[331832]: 2025-11-27 10:16:44.75159539 +0000 UTC m=+0.108230829 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, container_name=ovn_metadata_agent, io.buildah.version=1.41.3, config_id=ovn_metadata_agent, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:16:44 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:16:44 localhost podman[331833]: 2025-11-27 10:16:44.836755057 +0000 UTC m=+0.191759152 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:16:44 localhost podman[331833]: 2025-11-27 10:16:44.85584091 +0000 UTC m=+0.210845015 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125) Nov 27 05:16:44 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:44 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 2.453674623115578e-06 of space, bias 1.0, pg target 0.00048828125 quantized to 32 (current 32) Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:16:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.00202537208612507 of space, bias 4.0, pg target 1.6121961805555556 quantized to 16 (current 16) Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:16:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:16:45 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:45 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e"]} : dispatch Nov 27 05:16:45 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth caps", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e"]} : dispatch Nov 27 05:16:45 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth caps", "entity": "client.bob", "caps": ["mds", "allow rw path=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177", "osd", "allow rw pool=manila_data namespace=fsvolumens_b798f36f-5684-4eeb-9973-3cbbea82d29e"]}]': finished Nov 27 05:16:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v692: 177 pgs: 177 active+clean; 221 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 163 KiB/s wr, 8 op/s Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.225 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.225 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.226 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.475 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.476 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.476 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.476 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.937 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.966 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.993 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:16:46 localhost nova_compute[284026]: 2025-11-27 10:16:46.993 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:16:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "Joe", "format": "json"}]: dispatch Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:Joe, format:json, prefix:fs subvolume deauthorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes WARNING volumes.fs.operations.versions.subvolume_v1] deauthorized called for already-removed authID 'Joe' for subvolume '542efcb9-ee89-45b6-928b-9c3f42f649e6' Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:Joe, format:json, prefix:fs subvolume deauthorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "Joe", "format": "json"}]: dispatch Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:Joe, format:json, prefix:fs subvolume evict, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=Joe, client_metadata.root=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:Joe, format:json, prefix:fs subvolume evict, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v693: 177 pgs: 177 active+clean; 221 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 436 B/s rd, 141 KiB/s wr, 9 op/s Nov 27 05:16:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "bob", "format": "json"}]: dispatch Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.bob", "format": "json"} v 0) Nov 27 05:16:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:47 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.bob"} v 0) Nov 27 05:16:47 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.bob"} : dispatch Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:bob, format:json, prefix:fs subvolume deauthorize, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "auth_id": "bob", "format": "json"}]: dispatch Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=bob, client_metadata.root=/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e/5148dd48-1131-4a96-9b35-d2e0d08c5177 Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:bob, format:json, prefix:fs subvolume evict, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e267 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:48 localhost nova_compute[284026]: 2025-11-27 10:16:48.232 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.bob", "format": "json"} : dispatch Nov 27 05:16:48 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.bob"} : dispatch Nov 27 05:16:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.bob"} : dispatch Nov 27 05:16:48 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.bob"}]': finished Nov 27 05:16:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:16:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/572530374' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:16:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:16:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/572530374' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:16:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v694: 177 pgs: 177 active+clean; 221 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 132 KiB/s wr, 8 op/s Nov 27 05:16:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "format": "json"}]: dispatch Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "tempest-cephx-id-471164033", "format": "json"}]: dispatch Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume deauthorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} v 0) Nov 27 05:16:50 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} : dispatch Nov 27 05:16:50 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.tempest-cephx-id-471164033"} v 0) Nov 27 05:16:50 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-471164033"} : dispatch Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume deauthorize, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "auth_id": "tempest-cephx-id-471164033", "format": "json"}]: dispatch Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume evict, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=tempest-cephx-id-471164033, client_metadata.root=/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6/94ea462b-70b9-4eba-9293-91dc55f3c4af Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:tempest-cephx-id-471164033, format:json, prefix:fs subvolume evict, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:16:50 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.tempest-cephx-id-471164033", "format": "json"} : dispatch Nov 27 05:16:50 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-471164033"} : dispatch Nov 27 05:16:50 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.tempest-cephx-id-471164033"} : dispatch Nov 27 05:16:50 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.tempest-cephx-id-471164033"}]': finished Nov 27 05:16:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "format": "json"}]: dispatch Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:7aeed949-6381-43a8-bac5-b37209c017f2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:7aeed949-6381-43a8-bac5-b37209c017f2, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:51 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7aeed949-6381-43a8-bac5-b37209c017f2' of type subvolume Nov 27 05:16:51 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:51.341+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '7aeed949-6381-43a8-bac5-b37209c017f2' of type subvolume Nov 27 05:16:51 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "7aeed949-6381-43a8-bac5-b37209c017f2", "force": true, "format": "json"}]: dispatch Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/7aeed949-6381-43a8-bac5-b37209c017f2'' moved to trashcan Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:51 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:7aeed949-6381-43a8-bac5-b37209c017f2, vol_name:cephfs) < "" Nov 27 05:16:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v695: 177 pgs: 177 active+clean; 222 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 132 KiB/s wr, 8 op/s Nov 27 05:16:51 localhost nova_compute[284026]: 2025-11-27 10:16:51.939 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e267 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:53 localhost nova_compute[284026]: 2025-11-27 10:16:53.233 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v696: 177 pgs: 177 active+clean; 222 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 132 KiB/s wr, 8 op/s Nov 27 05:16:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3b393b3f-0cd7-4dce-b7ff-34036516ff96", "format": "json"}]: dispatch Nov 27 05:16:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:16:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:16:53 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:16:53 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "auth_id": "Joe", "format": "json"}]: dispatch Nov 27 05:16:53 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:Joe, format:json, prefix:fs subvolume deauthorize, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:53 localhost podman[331871]: 2025-11-27 10:16:53.995559083 +0000 UTC m=+0.088702384 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:16:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.Joe", "format": "json"} v 0) Nov 27 05:16:54 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:54 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.Joe"} v 0) Nov 27 05:16:54 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.Joe"} : dispatch Nov 27 05:16:54 localhost podman[331873]: 2025-11-27 10:16:54.056010936 +0000 UTC m=+0.138300806 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.openshift.expose-services=, version=9.6, architecture=x86_64, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, config_id=edpm, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., container_name=openstack_network_exporter, managed_by=edpm_ansible, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., vcs-type=git, com.redhat.component=ubi9-minimal-container, build-date=2025-08-20T13:12:41, io.buildah.version=1.33.7, distribution-scope=public) Nov 27 05:16:54 localhost podman[331873]: 2025-11-27 10:16:54.074037481 +0000 UTC m=+0.156327351 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.33.7, vcs-type=git, vendor=Red Hat, Inc., version=9.6, maintainer=Red Hat, Inc., url=https://catalog.redhat.com/en/search?searchType=containers, managed_by=edpm_ansible, com.redhat.component=ubi9-minimal-container, config_id=edpm, build-date=2025-08-20T13:12:41, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., release=1755695350, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.tags=minimal rhel9, architecture=x86_64, container_name=openstack_network_exporter, io.openshift.expose-services=, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal) Nov 27 05:16:54 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:Joe, format:json, prefix:fs subvolume deauthorize, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "auth_id": "Joe", "format": "json"}]: dispatch Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:Joe, format:json, prefix:fs subvolume evict, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=Joe, client_metadata.root=/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a/f0524ce0-0688-43a4-8714-4f847207d982 Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:Joe, format:json, prefix:fs subvolume evict, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:16:54 localhost systemd[1]: tmp-crun.hDM53F.mount: Deactivated successfully. Nov 27 05:16:54 localhost podman[331872]: 2025-11-27 10:16:54.165658611 +0000 UTC m=+0.251098955 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:16:54 localhost podman[331871]: 2025-11-27 10:16:54.183753778 +0000 UTC m=+0.276897059 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_id=edpm, container_name=ceilometer_agent_compute, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:16:54 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:16:54 localhost podman[331872]: 2025-11-27 10:16:54.205223874 +0000 UTC m=+0.290664168 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:16:54 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:16:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "format": "json"}]: dispatch Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:54.644+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b798f36f-5684-4eeb-9973-3cbbea82d29e' of type subvolume Nov 27 05:16:54 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'b798f36f-5684-4eeb-9973-3cbbea82d29e' of type subvolume Nov 27 05:16:54 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "b798f36f-5684-4eeb-9973-3cbbea82d29e", "force": true, "format": "json"}]: dispatch Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/b798f36f-5684-4eeb-9973-3cbbea82d29e'' moved to trashcan Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:16:54 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:b798f36f-5684-4eeb-9973-3cbbea82d29e, vol_name:cephfs) < "" Nov 27 05:16:54 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.Joe", "format": "json"} : dispatch Nov 27 05:16:54 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.Joe"} : dispatch Nov 27 05:16:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.Joe"} : dispatch Nov 27 05:16:54 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.Joe"}]': finished Nov 27 05:16:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v697: 177 pgs: 177 active+clean; 223 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 262 B/s rd, 186 KiB/s wr, 10 op/s Nov 27 05:16:55 localhost openstack_network_exporter[244641]: ERROR 10:16:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:16:55 localhost openstack_network_exporter[244641]: ERROR 10:16:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:16:55 localhost openstack_network_exporter[244641]: Nov 27 05:16:55 localhost openstack_network_exporter[244641]: ERROR 10:16:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:16:55 localhost openstack_network_exporter[244641]: ERROR 10:16:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:16:55 localhost openstack_network_exporter[244641]: ERROR 10:16:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:16:55 localhost openstack_network_exporter[244641]: Nov 27 05:16:56 localhost nova_compute[284026]: 2025-11-27 10:16:56.941 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "f0f022e4-d548-4f6e-8f19-eb0f75b94c6a", "format": "json"}]: dispatch Nov 27 05:16:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:16:57 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "auth_id": "admin", "tenant_id": "e5da946dcab84022ab5dfc2c601c61e9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:16:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:admin, format:json, prefix:fs subvolume authorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:16:57 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin", "format": "json"} v 0) Nov 27 05:16:57 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin", "format": "json"} : dispatch Nov 27 05:16:57 localhost ceph-mgr[290377]: [volumes ERROR volumes.fs.operations.versions.subvolume_v1] auth ID: admin exists and not created by mgr plugin. Not allowed to modify Nov 27 05:16:57 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:admin, format:json, prefix:fs subvolume authorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:16:57 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:16:57.229+0000 7f54df954640 -1 mgr.server reply reply (1) Operation not permitted auth ID: admin exists and not created by mgr plugin. Not allowed to modify Nov 27 05:16:57 localhost ceph-mgr[290377]: mgr.server reply reply (1) Operation not permitted auth ID: admin exists and not created by mgr plugin. Not allowed to modify Nov 27 05:16:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v698: 177 pgs: 177 active+clean; 223 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 118 KiB/s wr, 8 op/s Nov 27 05:16:57 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin", "format": "json"} : dispatch Nov 27 05:16:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e267 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:16:58 localhost nova_compute[284026]: 2025-11-27 10:16:58.236 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:16:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v699: 177 pgs: 177 active+clean; 223 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 117 KiB/s wr, 7 op/s Nov 27 05:17:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v700: 177 pgs: 177 active+clean; 223 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 143 KiB/s wr, 8 op/s Nov 27 05:17:01 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "auth_id": "david", "tenant_id": "e5da946dcab84022ab5dfc2c601c61e9", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:17:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:david, format:json, prefix:fs subvolume authorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:17:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.david", "format": "json"} v 0) Nov 27 05:17:01 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:01 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: Creating meta for ID david with tenant e5da946dcab84022ab5dfc2c601c61e9 Nov 27 05:17:01 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get-or-create", "entity": "client.david", "caps": ["mds", "allow rw path=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1", "osd", "allow rw pool=manila_data namespace=fsvolumens_f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "mon", "allow r"], "format": "json"} v 0) Nov 27 05:17:01 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.david", "caps": ["mds", "allow rw path=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1", "osd", "allow rw pool=manila_data namespace=fsvolumens_f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:17:01 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:david, format:json, prefix:fs subvolume authorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, tenant_id:e5da946dcab84022ab5dfc2c601c61e9, vol_name:cephfs) < "" Nov 27 05:17:01 localhost nova_compute[284026]: 2025-11-27 10:17:01.943 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:01 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:01 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.david", "caps": ["mds", "allow rw path=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1", "osd", "allow rw pool=manila_data namespace=fsvolumens_f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:17:01 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get-or-create", "entity": "client.david", "caps": ["mds", "allow rw path=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1", "osd", "allow rw pool=manila_data namespace=fsvolumens_f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "mon", "allow r"], "format": "json"} : dispatch Nov 27 05:17:01 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth get-or-create", "entity": "client.david", "caps": ["mds", "allow rw path=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1", "osd", "allow rw pool=manila_data namespace=fsvolumens_f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "mon", "allow r"], "format": "json"}]': finished Nov 27 05:17:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "f0f022e4-d548-4f6e-8f19-eb0f75b94c6a_b97a3364-d064-4493-9214-3dbd881d85fb", "force": true, "format": "json"}]: dispatch Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a_b97a3364-d064-4493-9214-3dbd881d85fb, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a_b97a3364-d064-4493-9214-3dbd881d85fb, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:02 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "f0f022e4-d548-4f6e-8f19-eb0f75b94c6a", "force": true, "format": "json"}]: dispatch Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:02 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:f0f022e4-d548-4f6e-8f19-eb0f75b94c6a, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e267 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:03 localhost nova_compute[284026]: 2025-11-27 10:17:03.238 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v701: 177 pgs: 177 active+clean; 223 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s rd, 98 KiB/s wr, 6 op/s Nov 27 05:17:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:17:03 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:17:03 localhost podman[331934]: 2025-11-27 10:17:03.989588012 +0000 UTC m=+0.087681176 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.build-date=20251125, tcib_managed=true, config_id=ovn_controller, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, container_name=ovn_controller, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS) Nov 27 05:17:04 localhost podman[331935]: 2025-11-27 10:17:04.072570641 +0000 UTC m=+0.167974303 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:17:04 localhost podman[331935]: 2025-11-27 10:17:04.081983784 +0000 UTC m=+0.177387436 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:17:04 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:17:04 localhost podman[331934]: 2025-11-27 10:17:04.133150888 +0000 UTC m=+0.231244072 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ovn_controller, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:17:04 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:17:04 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume create", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "size": 1073741824, "namespace_isolated": true, "mode": "0755", "format": "json"}]: dispatch Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/0bf16750-11a5-4741-ab53-2faa45d38d7b/.meta.tmp' Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/0bf16750-11a5-4741-ab53-2faa45d38d7b/.meta.tmp' to config b'/volumes/_nogroup/0bf16750-11a5-4741-ab53-2faa45d38d7b/.meta' Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_create(format:json, mode:0755, namespace_isolated:True, prefix:fs subvolume create, size:1073741824, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:04 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume getpath", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "format": "json"}]: dispatch Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:04 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_getpath(format:json, prefix:fs subvolume getpath, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:05 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3ab9fc96-8ff8-4f79-9f27-173775cef137", "format": "json"}]: dispatch Nov 27 05:17:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v702: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 597 B/s rd, 142 KiB/s wr, 8 op/s Nov 27 05:17:05 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:06 localhost nova_compute[284026]: 2025-11-27 10:17:06.947 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v703: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 72 KiB/s wr, 6 op/s Nov 27 05:17:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e268 e268: 6 total, 6 up, 6 in Nov 27 05:17:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e268 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:08 localhost nova_compute[284026]: 2025-11-27 10:17:08.241 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:08 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume authorize", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "auth_id": "david", "tenant_id": "31016ff904ec4126b31d88584782c763", "access_level": "rw", "format": "json"}]: dispatch Nov 27 05:17:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_authorize(access_level:rw, auth_id:david, format:json, prefix:fs subvolume authorize, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:17:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.david", "format": "json"} v 0) Nov 27 05:17:08 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:08 localhost ceph-mgr[290377]: [volumes ERROR volumes.fs.operations.versions.subvolume_v1] auth ID: david is already in use Nov 27 05:17:08 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_authorize(access_level:rw, auth_id:david, format:json, prefix:fs subvolume authorize, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, tenant_id:31016ff904ec4126b31d88584782c763, vol_name:cephfs) < "" Nov 27 05:17:08 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:08.289+0000 7f54df954640 -1 mgr.server reply reply (1) Operation not permitted auth ID: david is already in use Nov 27 05:17:08 localhost ceph-mgr[290377]: mgr.server reply reply (1) Operation not permitted auth ID: david is already in use Nov 27 05:17:08 localhost podman[242678]: time="2025-11-27T10:17:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:17:08 localhost podman[242678]: @ - - [27/Nov/2025:10:17:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:17:08 localhost podman[242678]: @ - - [27/Nov/2025:10:17:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19755 "" "Go-http-client/1.1" Nov 27 05:17:09 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v705: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 85 KiB/s wr, 5 op/s Nov 27 05:17:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3ab9fc96-8ff8-4f79-9f27-173775cef137_07d9cfec-be14-46c2-a944-d74d07b916f8", "force": true, "format": "json"}]: dispatch Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137_07d9cfec-be14-46c2-a944-d74d07b916f8, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137_07d9cfec-be14-46c2-a944-d74d07b916f8, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:10 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3ab9fc96-8ff8-4f79-9f27-173775cef137", "force": true, "format": "json"}]: dispatch Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:10 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3ab9fc96-8ff8-4f79-9f27-173775cef137, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:11 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "auth_id": "david", "format": "json"}]: dispatch Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:david, format:json, prefix:fs subvolume deauthorize, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v706: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 81 KiB/s wr, 5 op/s Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes WARNING volumes.fs.operations.versions.subvolume_v1] deauthorized called for already-removed authID 'david' for subvolume '0bf16750-11a5-4741-ab53-2faa45d38d7b' Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:david, format:json, prefix:fs subvolume deauthorize, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:11 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "auth_id": "david", "format": "json"}]: dispatch Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:david, format:json, prefix:fs subvolume evict, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=david, client_metadata.root=/volumes/_nogroup/0bf16750-11a5-4741-ab53-2faa45d38d7b/bf078efb-3961-4c51-ae4d-00eb3f1e8592 Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:17:11 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:david, format:json, prefix:fs subvolume evict, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:11 localhost nova_compute[284026]: 2025-11-27 10:17:11.949 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e268 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:13 localhost nova_compute[284026]: 2025-11-27 10:17:13.244 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v707: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 81 KiB/s wr, 5 op/s Nov 27 05:17:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e269 e269: 6 total, 6 up, 6 in Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:14 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "1d66ba40-29b7-4fa4-9b4c-c7d562d68515", "format": "json"}]: dispatch Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:14 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e270 e270: 6 total, 6 up, 6 in Nov 27 05:17:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:17:14 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:17:15 localhost systemd[1]: tmp-crun.BQzqow.mount: Deactivated successfully. Nov 27 05:17:15 localhost podman[331982]: 2025-11-27 10:17:15.015945274 +0000 UTC m=+0.100265764 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, container_name=multipathd, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS) Nov 27 05:17:15 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "auth_id": "david", "format": "json"}]: dispatch Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:david, format:json, prefix:fs subvolume deauthorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:15 localhost podman[331981]: 2025-11-27 10:17:15.062624109 +0000 UTC m=+0.149072006 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_managed=true) Nov 27 05:17:15 localhost podman[331982]: 2025-11-27 10:17:15.082252876 +0000 UTC m=+0.166573426 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, config_id=multipathd, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, container_name=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true) Nov 27 05:17:15 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:17:15 localhost podman[331981]: 2025-11-27 10:17:15.099033777 +0000 UTC m=+0.185481644 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, tcib_managed=true) Nov 27 05:17:15 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:17:15 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.david", "format": "json"} v 0) Nov 27 05:17:15 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:15 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth rm", "entity": "client.david"} v 0) Nov 27 05:17:15 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.david"} : dispatch Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:david, format:json, prefix:fs subvolume deauthorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:15 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume evict", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "auth_id": "david", "format": "json"}]: dispatch Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_evict(auth_id:david, format:json, prefix:fs subvolume evict, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict clients with auth_name=david, client_metadata.root=/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7/c12cbeaf-e500-4e76-a47c-3c1dc7bb50d1 Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_v1] evict: joined all Nov 27 05:17:15 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_evict(auth_id:david, format:json, prefix:fs subvolume evict, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v710: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 276 B/s rd, 115 KiB/s wr, 5 op/s Nov 27 05:17:15 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.david", "format": "json"} : dispatch Nov 27 05:17:15 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.david"} : dispatch Nov 27 05:17:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth rm", "entity": "client.david"} : dispatch Nov 27 05:17:15 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' cmd='[{"prefix": "auth rm", "entity": "client.david"}]': finished Nov 27 05:17:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:17:16 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:17:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:17:16 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:17:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:17:16 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 40075cf7-fc66-446a-94ba-345225bb4655 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:17:16 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 40075cf7-fc66-446a-94ba-345225bb4655 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:17:16 localhost ceph-mgr[290377]: [progress INFO root] Completed event 40075cf7-fc66-446a-94ba-345225bb4655 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:17:16 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:17:16 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:17:16 localhost nova_compute[284026]: 2025-11-27 10:17:16.951 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:17 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:17:17 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:17:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v711: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 107 KiB/s wr, 6 op/s Nov 27 05:17:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e270 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #52. Immutable memtables: 0. Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.189062) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 29] Flushing memtable with next log file: 52 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638189101, "job": 29, "event": "flush_started", "num_memtables": 1, "num_entries": 2761, "num_deletes": 255, "total_data_size": 3298874, "memory_usage": 3359424, "flush_reason": "Manual Compaction"} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 29] Level-0 flush table #53: started Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638203675, "cf_name": "default", "job": 29, "event": "table_file_creation", "file_number": 53, "file_size": 2116661, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 32009, "largest_seqno": 34764, "table_properties": {"data_size": 2105936, "index_size": 6591, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3077, "raw_key_size": 26989, "raw_average_key_size": 22, "raw_value_size": 2082782, "raw_average_value_size": 1712, "num_data_blocks": 286, "num_entries": 1216, "num_filter_entries": 1216, "num_deletions": 255, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238505, "oldest_key_time": 1764238505, "file_creation_time": 1764238638, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 53, "seqno_to_time_mapping": "N/A"}} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 29] Flush lasted 14670 microseconds, and 6559 cpu microseconds. Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.203728) [db/flush_job.cc:967] [default] [JOB 29] Level-0 flush table #53: 2116661 bytes OK Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.203756) [db/memtable_list.cc:519] [default] Level-0 commit table #53 started Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.209994) [db/memtable_list.cc:722] [default] Level-0 commit table #53: memtable #1 done Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.210016) EVENT_LOG_v1 {"time_micros": 1764238638210009, "job": 29, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.210039) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 29] Try to delete WAL files size 3285760, prev total WAL file size 3285760, number of live WAL files 2. Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000049.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.211066) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '7061786F73003133303532' seq:72057594037927935, type:22 .. '7061786F73003133333034' seq:0, type:0; will stop at (end) Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 30] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 29 Base level 0, inputs: [53(2067KB)], [51(20MB)] Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638211147, "job": 30, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [53], "files_L6": [51], "score": -1, "input_data_size": 23478520, "oldest_snapshot_seqno": -1} Nov 27 05:17:18 localhost nova_compute[284026]: 2025-11-27 10:17:18.247 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 30] Generated table #54: 14807 keys, 21838986 bytes, temperature: kUnknown Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638401737, "cf_name": "default", "job": 30, "event": "table_file_creation", "file_number": 54, "file_size": 21838986, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 21750116, "index_size": 50753, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 37061, "raw_key_size": 394210, "raw_average_key_size": 26, "raw_value_size": 21494769, "raw_average_value_size": 1451, "num_data_blocks": 1919, "num_entries": 14807, "num_filter_entries": 14807, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238638, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 54, "seqno_to_time_mapping": "N/A"}} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.402328) [db/compaction/compaction_job.cc:1663] [default] [JOB 30] Compacted 1@0 + 1@6 files to L6 => 21838986 bytes Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.403934) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 123.0 rd, 114.4 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(2.0, 20.4 +0.0 blob) out(20.8 +0.0 blob), read-write-amplify(21.4) write-amplify(10.3) OK, records in: 15341, records dropped: 534 output_compression: NoCompression Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.403965) EVENT_LOG_v1 {"time_micros": 1764238638403951, "job": 30, "event": "compaction_finished", "compaction_time_micros": 190878, "compaction_time_cpu_micros": 68066, "output_level": 6, "num_output_files": 1, "total_output_size": 21838986, "num_input_records": 15341, "num_output_records": 14807, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000053.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638404573, "job": 30, "event": "table_file_deletion", "file_number": 53} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000051.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238638407778, "job": 30, "event": "table_file_deletion", "file_number": 51} Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.210915) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.407967) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.407975) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.407978) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.407981) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:18.407985) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:18 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "1d66ba40-29b7-4fa4-9b4c-c7d562d68515_8a2cd4bb-da43-4aad-92da-be0a7a5517c0", "force": true, "format": "json"}]: dispatch Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515_8a2cd4bb-da43-4aad-92da-be0a7a5517c0, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515_8a2cd4bb-da43-4aad-92da-be0a7a5517c0, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:18 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "1d66ba40-29b7-4fa4-9b4c-c7d562d68515", "force": true, "format": "json"}]: dispatch Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:18 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:1d66ba40-29b7-4fa4-9b4c-c7d562d68515, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e271 e271: 6 total, 6 up, 6 in Nov 27 05:17:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "format": "json"}]: dispatch Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:19 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:19.086+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '0bf16750-11a5-4741-ab53-2faa45d38d7b' of type subvolume Nov 27 05:17:19 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '0bf16750-11a5-4741-ab53-2faa45d38d7b' of type subvolume Nov 27 05:17:19 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "0bf16750-11a5-4741-ab53-2faa45d38d7b", "force": true, "format": "json"}]: dispatch Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/0bf16750-11a5-4741-ab53-2faa45d38d7b'' moved to trashcan Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:17:19 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:0bf16750-11a5-4741-ab53-2faa45d38d7b, vol_name:cephfs) < "" Nov 27 05:17:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v713: 177 pgs: 177 active+clean; 224 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 341 B/s rd, 98 KiB/s wr, 6 op/s Nov 27 05:17:20 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:17:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:17:20 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:17:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v714: 177 pgs: 177 active+clean; 225 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 400 B/s rd, 145 KiB/s wr, 8 op/s Nov 27 05:17:21 localhost nova_compute[284026]: 2025-11-27 10:17:21.954 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "e25efe8a-e740-4099-a606-46633283031b", "format": "json"}]: dispatch Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e25efe8a-e740-4099-a606-46633283031b, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:e25efe8a-e740-4099-a606-46633283031b, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "format": "json"}]: dispatch Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:22 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:22.615+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '542efcb9-ee89-45b6-928b-9c3f42f649e6' of type subvolume Nov 27 05:17:22 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '542efcb9-ee89-45b6-928b-9c3f42f649e6' of type subvolume Nov 27 05:17:22 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "542efcb9-ee89-45b6-928b-9c3f42f649e6", "force": true, "format": "json"}]: dispatch Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/542efcb9-ee89-45b6-928b-9c3f42f649e6'' moved to trashcan Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:17:22 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:542efcb9-ee89-45b6-928b-9c3f42f649e6, vol_name:cephfs) < "" Nov 27 05:17:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e271 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:23 localhost nova_compute[284026]: 2025-11-27 10:17:23.272 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e272 e272: 6 total, 6 up, 6 in Nov 27 05:17:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v716: 177 pgs: 177 active+clean; 225 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 127 B/s rd, 67 KiB/s wr, 3 op/s Nov 27 05:17:24 localhost nova_compute[284026]: 2025-11-27 10:17:24.001 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:24.002 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=27, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=26) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:17:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:24.004 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 10 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:17:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:17:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:17:24 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:17:25 localhost podman[332103]: 2025-11-27 10:17:25.008162326 +0000 UTC m=+0.096445401 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:17:25 localhost podman[332104]: 2025-11-27 10:17:25.057705467 +0000 UTC m=+0.141677487 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:17:25 localhost podman[332104]: 2025-11-27 10:17:25.06974543 +0000 UTC m=+0.153717430 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:17:25 localhost podman[332103]: 2025-11-27 10:17:25.077686934 +0000 UTC m=+0.165970059 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2) Nov 27 05:17:25 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:17:25 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:17:25 localhost podman[332105]: 2025-11-27 10:17:25.163927591 +0000 UTC m=+0.246741809 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., release=1755695350, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, maintainer=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, architecture=x86_64, vcs-type=git, config_id=edpm, com.redhat.component=ubi9-minimal-container, build-date=2025-08-20T13:12:41, name=ubi9-minimal, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, distribution-scope=public, url=https://catalog.redhat.com/en/search?searchType=containers, io.buildah.version=1.33.7, managed_by=edpm_ansible, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, version=9.6, container_name=openstack_network_exporter) Nov 27 05:17:25 localhost podman[332105]: 2025-11-27 10:17:25.210965593 +0000 UTC m=+0.293779811 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, distribution-scope=public, managed_by=edpm_ansible, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, io.openshift.expose-services=, maintainer=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, release=1755695350, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., build-date=2025-08-20T13:12:41, name=ubi9-minimal, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vendor=Red Hat, Inc., architecture=x86_64, container_name=openstack_network_exporter, version=9.6, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, io.openshift.tags=minimal rhel9, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git) Nov 27 05:17:25 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:17:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v717: 177 pgs: 177 active+clean; 225 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 639 B/s rd, 106 KiB/s wr, 5 op/s Nov 27 05:17:25 localhost openstack_network_exporter[244641]: ERROR 10:17:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:17:25 localhost openstack_network_exporter[244641]: ERROR 10:17:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:17:25 localhost openstack_network_exporter[244641]: ERROR 10:17:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:17:25 localhost openstack_network_exporter[244641]: ERROR 10:17:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:17:25 localhost openstack_network_exporter[244641]: Nov 27 05:17:25 localhost openstack_network_exporter[244641]: ERROR 10:17:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:17:25 localhost openstack_network_exporter[244641]: Nov 27 05:17:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "15efeb80-6730-4067-836c-71b94433ea7a", "format": "json"}]: dispatch Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:15efeb80-6730-4067-836c-71b94433ea7a, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:15efeb80-6730-4067-836c-71b94433ea7a, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:26 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:26.756+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '15efeb80-6730-4067-836c-71b94433ea7a' of type subvolume Nov 27 05:17:26 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '15efeb80-6730-4067-836c-71b94433ea7a' of type subvolume Nov 27 05:17:26 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "15efeb80-6730-4067-836c-71b94433ea7a", "force": true, "format": "json"}]: dispatch Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/15efeb80-6730-4067-836c-71b94433ea7a'' moved to trashcan Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:17:26 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:15efeb80-6730-4067-836c-71b94433ea7a, vol_name:cephfs) < "" Nov 27 05:17:26 localhost nova_compute[284026]: 2025-11-27 10:17:26.955 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:27 localhost nova_compute[284026]: 2025-11-27 10:17:27.483 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_running_deleted_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v718: 177 pgs: 177 active+clean; 225 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 590 B/s rd, 98 KiB/s wr, 5 op/s Nov 27 05:17:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e272 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:28 localhost nova_compute[284026]: 2025-11-27 10:17:28.302 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "e25efe8a-e740-4099-a606-46633283031b_2159b8e0-0d97-41e6-b105-929262221134", "force": true, "format": "json"}]: dispatch Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e25efe8a-e740-4099-a606-46633283031b_2159b8e0-0d97-41e6-b105-929262221134, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e25efe8a-e740-4099-a606-46633283031b_2159b8e0-0d97-41e6-b105-929262221134, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:28 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "e25efe8a-e740-4099-a606-46633283031b", "force": true, "format": "json"}]: dispatch Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e25efe8a-e740-4099-a606-46633283031b, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:28 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:e25efe8a-e740-4099-a606-46633283031b, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e273 e273: 6 total, 6 up, 6 in Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #55. Immutable memtables: 0. Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.848831) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 31] Flushing memtable with next log file: 55 Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238648848932, "job": 31, "event": "flush_started", "num_memtables": 1, "num_entries": 452, "num_deletes": 253, "total_data_size": 348363, "memory_usage": 357416, "flush_reason": "Manual Compaction"} Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 31] Level-0 flush table #56: started Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238648853658, "cf_name": "default", "job": 31, "event": "table_file_creation", "file_number": 56, "file_size": 226141, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 34769, "largest_seqno": 35216, "table_properties": {"data_size": 223653, "index_size": 535, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 901, "raw_key_size": 7039, "raw_average_key_size": 21, "raw_value_size": 218425, "raw_average_value_size": 652, "num_data_blocks": 24, "num_entries": 335, "num_filter_entries": 335, "num_deletions": 253, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238638, "oldest_key_time": 1764238638, "file_creation_time": 1764238648, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 56, "seqno_to_time_mapping": "N/A"}} Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 31] Flush lasted 4885 microseconds, and 1753 cpu microseconds. Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.853729) [db/flush_job.cc:967] [default] [JOB 31] Level-0 flush table #56: 226141 bytes OK Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.853751) [db/memtable_list.cc:519] [default] Level-0 commit table #56 started Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.856220) [db/memtable_list.cc:722] [default] Level-0 commit table #56: memtable #1 done Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.856241) EVENT_LOG_v1 {"time_micros": 1764238648856235, "job": 31, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.856262) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 31] Try to delete WAL files size 345510, prev total WAL file size 345510, number of live WAL files 2. Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000052.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.857080) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6D6772737461740034323535' seq:72057594037927935, type:22 .. '6D6772737461740034353037' seq:0, type:0; will stop at (end) Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 32] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 31 Base level 0, inputs: [56(220KB)], [54(20MB)] Nov 27 05:17:28 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238648857132, "job": 32, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [56], "files_L6": [54], "score": -1, "input_data_size": 22065127, "oldest_snapshot_seqno": -1} Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 32] Generated table #57: 14618 keys, 19937612 bytes, temperature: kUnknown Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238649010196, "cf_name": "default", "job": 32, "event": "table_file_creation", "file_number": 57, "file_size": 19937612, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 19854646, "index_size": 45297, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 36613, "raw_key_size": 390558, "raw_average_key_size": 26, "raw_value_size": 19607154, "raw_average_value_size": 1341, "num_data_blocks": 1691, "num_entries": 14618, "num_filter_entries": 14618, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238648, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 57, "seqno_to_time_mapping": "N/A"}} Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.010572) [db/compaction/compaction_job.cc:1663] [default] [JOB 32] Compacted 1@0 + 1@6 files to L6 => 19937612 bytes Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.012629) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 144.0 rd, 130.2 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.2, 20.8 +0.0 blob) out(19.0 +0.0 blob), read-write-amplify(185.7) write-amplify(88.2) OK, records in: 15142, records dropped: 524 output_compression: NoCompression Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.012681) EVENT_LOG_v1 {"time_micros": 1764238649012662, "job": 32, "event": "compaction_finished", "compaction_time_micros": 153178, "compaction_time_cpu_micros": 53845, "output_level": 6, "num_output_files": 1, "total_output_size": 19937612, "num_input_records": 15142, "num_output_records": 14618, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000056.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238649012968, "job": 32, "event": "table_file_deletion", "file_number": 56} Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000054.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238649016038, "job": 32, "event": "table_file_deletion", "file_number": 54} Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:28.856966) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.016106) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.016113) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.016116) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.016120) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:17:29.016419) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:17:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v720: 177 pgs: 177 active+clean; 225 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 41 KiB/s wr, 3 op/s Nov 27 05:17:30 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume deauthorize", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "auth_id": "admin", "format": "json"}]: dispatch Nov 27 05:17:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_deauthorize(auth_id:admin, format:json, prefix:fs subvolume deauthorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:30 localhost ceph-mgr[290377]: [volumes ERROR volumes.fs.operations.versions.subvolume_v1] auth ID: admin doesn't exist Nov 27 05:17:30 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_deauthorize(auth_id:admin, format:json, prefix:fs subvolume deauthorize, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:30 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:30.950+0000 7f54df954640 -1 mgr.server reply reply (2) No such file or directory auth ID: admin doesn't exist Nov 27 05:17:30 localhost ceph-mgr[290377]: mgr.server reply reply (2) No such file or directory auth ID: admin doesn't exist Nov 27 05:17:31 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "format": "json"}]: dispatch Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:31 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:31.146+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7' of type subvolume Nov 27 05:17:31 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume 'f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7' of type subvolume Nov 27 05:17:31 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7", "force": true, "format": "json"}]: dispatch Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7'' moved to trashcan Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:17:31 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:f5e6db00-6f8b-4289-b33d-1ecd2dc9a6f7, vol_name:cephfs) < "" Nov 27 05:17:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v721: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 750 B/s rd, 105 KiB/s wr, 6 op/s Nov 27 05:17:31 localhost nova_compute[284026]: 2025-11-27 10:17:31.924 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:31 localhost nova_compute[284026]: 2025-11-27 10:17:31.925 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:31 localhost nova_compute[284026]: 2025-11-27 10:17:31.957 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:32 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e274 e274: 6 total, 6 up, 6 in Nov 27 05:17:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e274 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:33 localhost nova_compute[284026]: 2025-11-27 10:17:33.305 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v723: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 67 KiB/s wr, 4 op/s Nov 27 05:17:33 localhost nova_compute[284026]: 2025-11-27 10:17:33.729 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:33 localhost nova_compute[284026]: 2025-11-27 10:17:33.969 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:34 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:34.006 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '27'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:17:34 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot create", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "6d094a69-74c8-40ec-a8a3-1b45125bad56", "format": "json"}]: dispatch Nov 27 05:17:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:34 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_create(format:json, prefix:fs subvolume snapshot create, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:34 localhost nova_compute[284026]: 2025-11-27 10:17:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:34 localhost nova_compute[284026]: 2025-11-27 10:17:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:34 localhost nova_compute[284026]: 2025-11-27 10:17:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:34 localhost nova_compute[284026]: 2025-11-27 10:17:34.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:17:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:17:34 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:17:35 localhost podman[332167]: 2025-11-27 10:17:35.027071786 +0000 UTC m=+0.125066869 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, container_name=ovn_controller, managed_by=edpm_ansible, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:17:35 localhost systemd[1]: tmp-crun.8gcAcr.mount: Deactivated successfully. Nov 27 05:17:35 localhost podman[332168]: 2025-11-27 10:17:35.113727395 +0000 UTC m=+0.206169369 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors ) Nov 27 05:17:35 localhost podman[332168]: 2025-11-27 10:17:35.12804341 +0000 UTC m=+0.220485424 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:17:35 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:17:35 localhost podman[332167]: 2025-11-27 10:17:35.152984959 +0000 UTC m=+0.250980052 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, container_name=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller) Nov 27 05:17:35 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:17:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v724: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 767 B/s rd, 101 KiB/s wr, 5 op/s Nov 27 05:17:36 localhost nova_compute[284026]: 2025-11-27 10:17:36.728 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:36 localhost nova_compute[284026]: 2025-11-27 10:17:36.959 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v725: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 708 B/s rd, 94 KiB/s wr, 5 op/s Nov 27 05:17:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e274 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.347 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:38 localhost podman[242678]: time="2025-11-27T10:17:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:17:38 localhost podman[242678]: @ - - [27/Nov/2025:10:17:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:17:38 localhost podman[242678]: @ - - [27/Nov/2025:10:17:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19747 "" "Go-http-client/1.1" Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.823 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.823 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.824 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:17:38 localhost nova_compute[284026]: 2025-11-27 10:17:38.825 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:17:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e275 e275: 6 total, 6 up, 6 in Nov 27 05:17:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:17:39 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3644445080' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:17:39 localhost nova_compute[284026]: 2025-11-27 10:17:39.265 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.440s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:17:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v727: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 35 KiB/s wr, 2 op/s Nov 27 05:17:39 localhost nova_compute[284026]: 2025-11-27 10:17:39.803 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:17:39 localhost nova_compute[284026]: 2025-11-27 10:17:39.803 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.043 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.045 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11037MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.046 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.046 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.823 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.824 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.894 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing inventories for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:804#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.952 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating ProviderTree inventory for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f from _refresh_and_get_inventory using data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} _refresh_and_get_inventory /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:768#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.953 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Updating inventory in ProviderTree for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f with inventory: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:176#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.969 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing aggregate associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, aggregates: None _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:813#033[00m Nov 27 05:17:40 localhost nova_compute[284026]: 2025-11-27 10:17:40.994 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Refreshing trait associations for resource provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f, traits: COMPUTE_IMAGE_TYPE_AKI,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NET_VIF_MODEL_PCNET,HW_CPU_X86_F16C,COMPUTE_NET_VIF_MODEL_E1000,COMPUTE_DEVICE_TAGGING,COMPUTE_STORAGE_BUS_FDC,COMPUTE_IMAGE_TYPE_RAW,COMPUTE_IMAGE_TYPE_QCOW2,COMPUTE_SOCKET_PCI_NUMA_AFFINITY,HW_CPU_X86_SHA,COMPUTE_STORAGE_BUS_USB,HW_CPU_X86_CLMUL,COMPUTE_SECURITY_TPM_1_2,COMPUTE_IMAGE_TYPE_ARI,COMPUTE_NET_ATTACH_INTERFACE_WITH_TAG,COMPUTE_VOLUME_MULTI_ATTACH,COMPUTE_NET_VIF_MODEL_VIRTIO,COMPUTE_VIOMMU_MODEL_INTEL,COMPUTE_SECURITY_TPM_2_0,COMPUTE_NODE,COMPUTE_ACCELERATORS,COMPUTE_VOLUME_EXTEND,COMPUTE_NET_VIF_MODEL_RTL8139,COMPUTE_STORAGE_BUS_IDE,COMPUTE_VIOMMU_MODEL_VIRTIO,HW_CPU_X86_SSE2,HW_CPU_X86_BMI,COMPUTE_GRAPHICS_MODEL_VIRTIO,COMPUTE_GRAPHICS_MODEL_NONE,COMPUTE_RESCUE_BFV,COMPUTE_NET_VIF_MODEL_NE2K_PCI,COMPUTE_IMAGE_TYPE_AMI,COMPUTE_VIOMMU_MODEL_AUTO,HW_CPU_X86_AESNI,COMPUTE_NET_VIF_MODEL_SPAPR_VLAN,HW_CPU_X86_SSE41,COMPUTE_GRAPHICS_MODEL_BOCHS,HW_CPU_X86_BMI2,HW_CPU_X86_SSE,COMPUTE_TRUSTED_CERTS,COMPUTE_VOLUME_ATTACH_WITH_TAG,HW_CPU_X86_SSE42,COMPUTE_GRAPHICS_MODEL_CIRRUS,HW_CPU_X86_SSE4A,COMPUTE_SECURITY_UEFI_SECURE_BOOT,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_VIF_MODEL_VMXNET3,HW_CPU_X86_MMX,HW_CPU_X86_AVX,COMPUTE_GRAPHICS_MODEL_VGA,HW_CPU_X86_ABM,COMPUTE_STORAGE_BUS_SCSI,HW_CPU_X86_SSSE3,HW_CPU_X86_AMD_SVM,HW_CPU_X86_SVM,COMPUTE_STORAGE_BUS_VIRTIO,COMPUTE_STORAGE_BUS_SATA,HW_CPU_X86_FMA3,COMPUTE_NET_VIF_MODEL_E1000E,HW_CPU_X86_AVX2 _refresh_associations /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:825#033[00m Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.033 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:17:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "6d094a69-74c8-40ec-a8a3-1b45125bad56_8e393c54-8994-4a17-8c86-336ccc04c024", "force": true, "format": "json"}]: dispatch Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56_8e393c54-8994-4a17-8c86-336ccc04c024, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56_8e393c54-8994-4a17-8c86-336ccc04c024, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:41 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "6d094a69-74c8-40ec-a8a3-1b45125bad56", "force": true, "format": "json"}]: dispatch Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:41 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:17:41 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/3922379090' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:17:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v728: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 477 B/s rd, 59 KiB/s wr, 3 op/s Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.503 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.471s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.511 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:17:41 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:6d094a69-74c8-40ec-a8a3-1b45125bad56, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.735 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.737 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.738 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.692s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:17:41 localhost nova_compute[284026]: 2025-11-27 10:17:41.961 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e275 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:43 localhost nova_compute[284026]: 2025-11-27 10:17:43.352 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v729: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 50 KiB/s wr, 3 op/s Nov 27 05:17:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:43.584 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:17:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:43.584 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:17:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:17:43.585 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.142 12 DEBUG ceilometer.compute.discovery [-] instance data: {'id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'name': 'test', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'os_type': 'hvm', 'architecture': 'x86_64', 'OS-EXT-SRV-ATTR:instance_name': 'instance-00000002', 'OS-EXT-SRV-ATTR:host': 'np0005537446.localdomain', 'OS-EXT-STS:vm_state': 'running', 'tenant_id': '02a308de13ca4518beaddd402ba93936', 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'hostId': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'status': 'active', 'metadata': {}} discover_libvirt_polling /usr/lib/python3.9/site-packages/ceilometer/compute/discovery.py:228 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.143 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes.delta in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.148 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'eccfd8ab-2419-492f-be7f-46f0ef539968', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.143976', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5042db18-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '3aebe27d6a7ebebd739cd7ab9206d035fb2f6ef8a1b30e4138871b7ce289a47a'}]}, 'timestamp': '2025-11-27 10:17:44.149633', '_unique_id': 'f8b4151f3774492388fdb13ccf7e0fba'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.151 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.152 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.incoming.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.152 12 INFO ceilometer.polling.manager [-] Polling pollster memory.usage in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.169 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/memory.usage volume: 51.67578125 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'c3db5951-5455-4499-8708-eab92eb0da86', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'memory.usage', 'counter_type': 'gauge', 'counter_unit': 'MB', 'counter_volume': 51.67578125, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:17:44.152894', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0}, 'message_id': '50460cac-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.357875375, 'message_signature': '4fc97dee1ccd11a717a2ec1ad2e11435c0186ae62193347161d672aee9fa4482'}]}, 'timestamp': '2025-11-27 10:17:44.170442', '_unique_id': '2417f66f38334dba9b814c571bd8af4d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.171 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.172 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.latency, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.173 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.183 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.184 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.usage volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '75355054-76a8-44de-b1ec-2fc86e2eef79', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.173146', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '50482c76-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': 'a2e3f4950ea575336019bdf575d739963396f2c9a9984ba27694cbc8ee36c736'}, {'source': 'openstack', 'counter_name': 'disk.device.usage', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.173146', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '50483fea-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': 'a6d833b686eb9b2288682c7a49ba58427be165831e39638760bab90e053946f1'}]}, 'timestamp': '2025-11-27 10:17:44.184843', '_unique_id': '332b9caa44dc4170acf9eaaad736eb2d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.185 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.187 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.drop in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.187 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a7391f9d-5fb6-40db-9c46-2b1c20991f2b', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.187232', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '5048af84-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': 'c9533470f364736f563b0859046401d5f728cf3ed93f3f3dbca8cb2f288d4ef8'}]}, 'timestamp': '2025-11-27 10:17:44.187782', '_unique_id': '387d021da11e4aa1a5356a64a6d73b73'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.188 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.189 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.190 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes volume: 9770 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bf04826b-3f4e-4135-857b-f292e6f83c62', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 9770, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.190066', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '50491dca-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': 'fe52d92b049555e7d16e46e7759ad4a036564013296eb1a8ebc73f68c59c9e40'}]}, 'timestamp': '2025-11-27 10:17:44.190573', '_unique_id': '7c1ec0f5c34c41fbb14d4ee64b93aed3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.191 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.192 12 DEBUG ceilometer.polling.manager [-] Skip pollster disk.device.iops, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.192 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.192 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets volume: 114 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '976953c6-8050-4423-9128-8b2e64489f3f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 114, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.192945', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '50498f44-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '17d46d9dbf6341ce9c6dbe9a859eb334379ddfd1cd89de2b54cbf6ae8fdfc238'}]}, 'timestamp': '2025-11-27 10:17:44.193449', '_unique_id': 'bedc4ffae8f54104bdc94ea2df409867'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.194 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.195 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.bytes in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 389120 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.223 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.bytes volume: 512 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'bd0c3541-fc1c-4709-8697-b5aa561bf655', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 389120, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.195714', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '504e28b0-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '6e7dfa83f89e30f4ecec499b521597af003b173f98fe17bc5ca351addeff0e22'}, {'source': 'openstack', 'counter_name': 'disk.device.write.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 512, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.195714', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '504e3ce2-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '53c0ea30463bc985c4505045c552abb8f075ecb02c7c4fdf2d202b97fce1a55f'}]}, 'timestamp': '2025-11-27 10:17:44.224119', '_unique_id': '798fc46dc9d441f3919c7db63055ae04'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.225 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.226 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.latency in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.226 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 1143885286 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.227 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.latency volume: 93387932 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'aad0f986-5175-475d-8e2c-93bfbc07bbfa', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1143885286, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.226576', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '504eb064-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '17924165e7946a15c6593b4bdbc8477e10c261e58db4d8c56eec69f7e8e5cfe5'}, {'source': 'openstack', 'counter_name': 'disk.device.read.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 93387932, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.226576', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '504ec0e0-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': 'ac268be2fa2ce4e6ceba9e4feb952048c330bd378ba551d12b156dcf300cef6a'}]}, 'timestamp': '2025-11-27 10:17:44.227451', '_unique_id': '6c0969a75cc34441af1d98fde226e053'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.228 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.229 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.229 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 1272 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.230 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.requests volume: 124 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0800d54d-7b8d-4bef-a67a-29cd02234c01', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1272, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.229756', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '504f2c74-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '90833fc983b246a638d62a2ee028c76e2b790dee7e438729e91778a30ad68e94'}, {'source': 'openstack', 'counter_name': 'disk.device.read.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 124, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.229756', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '504f3d2c-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '5c5ec3a2f78cd74ffa29652c4ad097d283ab205f5aa01f2f82243b2306af2869'}]}, 'timestamp': '2025-11-27 10:17:44.230739', '_unique_id': '1b79afb1c1fc45f983d409efb52d9e81'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.231 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.233 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.drop in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.233 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.drop volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '1690e9a4-30ea-42db-86be-6edaae94d6d8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.drop', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.233185', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '504fb554-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '58794911b3c453c0b93801f86c9aa0798352cfa55a14f04ec218ca3443d72c7f'}]}, 'timestamp': '2025-11-27 10:17:44.233851', '_unique_id': '8dcc42ecad584db4a39a7a96de4a495d'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.234 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.235 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.bytes in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.236 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.bytes volume: 6875 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '51783832-b632-432d-b9bf-52cac096dfdb', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 6875, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.236103', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '505024bc-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '0f60275d8db8d4006395771b7a1728219920c182283ab2984e64267255f269f2'}]}, 'timestamp': '2025-11-27 10:17:44.236663', '_unique_id': '7243310c8a454c14b88dd29ae4bccaa7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.237 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.238 12 DEBUG ceilometer.polling.manager [-] Skip pollster network.outgoing.bytes.rate, no new resources found this cycle poll_and_notify /usr/lib/python3.9/site-packages/ceilometer/polling/manager.py:193 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.238 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.latency in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 1134280252 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.239 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.latency volume: 28338283 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '02edc657-9ef8-40e7-bb03-bc635bd6b302', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 1134280252, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.238968', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '505093ca-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '7c4a6ceaf61eea7f5f620d27a046cdb062e9328f6bdf87847a6d6064ed07145f'}, {'source': 'openstack', 'counter_name': 'disk.device.write.latency', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 28338283, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.238968', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5050a568-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '91c7c13eb9a86de390506a6442282a9e31db6a7fcedf513221ba90143eb2137d'}]}, 'timestamp': '2025-11-27 10:17:44.239857', '_unique_id': 'c42ce023eded40f2aaa2d582e8203c97'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.240 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.241 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.242 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets volume: 61 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'ff9d3c71-10c6-44a6-855f-9a70decf05e7', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 61, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.242076', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '50510d78-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '26ce60f6c03a31050dd19fd23d8aee1b07bdc25874b2d383bbcbcb63c63e4a14'}]}, 'timestamp': '2025-11-27 10:17:44.242578', '_unique_id': '29df340d79f04a13b40bd44d77950928'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.243 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.244 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.packets.error in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.244 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a7d6959c-43f5-4aca-90f0-7d0e79935720', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.244725', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '505174fc-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '74589839eb97fc3f00df105d7f3ac0546d5a08fc108cbb0d998574b64871451a'}]}, 'timestamp': '2025-11-27 10:17:44.245199', '_unique_id': '8a28094c0aac4ffcaa6571831f5e5f47'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.246 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.247 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.bytes in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 35560448 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.247 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.read.bytes volume: 2154496 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '8648f005-f7fe-4237-a532-a1d9eb407402', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 35560448, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.247334', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5051dbae-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '398681b9dd1df02fc668b25b05faacf1d9e008ace6c26d0c33af647c8cf248fe'}, {'source': 'openstack', 'counter_name': 'disk.device.read.bytes', 'counter_type': 'cumulative', 'counter_unit': 'B', 'counter_volume': 2154496, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.247334', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5051eca2-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '9bac53a09eb64bc4a165e1fe323a34222af4735f80622acab5562849cae8add6'}]}, 'timestamp': '2025-11-27 10:17:44.248233', '_unique_id': 'fa0c6ae788044d5eb615dfd708f00cdb'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.249 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.250 12 INFO ceilometer.polling.manager [-] Polling pollster network.outgoing.bytes.delta in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.250 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.outgoing.bytes.delta volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '029882e7-dbe3-4231-89af-4587513e98cf', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.outgoing.bytes.delta', 'counter_type': 'delta', 'counter_unit': 'B', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.250417', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '50525494-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '4c190a3ef2899d32be435eb2be35113be49a11e23131f02f18c1cffe5118994c'}]}, 'timestamp': '2025-11-27 10:17:44.250925', '_unique_id': '1c864b28a0754104a884c8584d3c2c37'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.251 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.252 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.capacity in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.253 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.capacity volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '3535767f-a3a2-46d4-a45a-060fe2516a9f', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.253092', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5052bbaa-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': 'e57b35d72d30b373848638050f32a4abe1191355cee08ad972e266d130d1d7b5'}, {'source': 'openstack', 'counter_name': 'disk.device.capacity', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.253092', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5052cd66-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': 'd86d987b036f0fa59efd9a68b684df69ec6fd4939fcbc6ef3211e730e25c3260'}]}, 'timestamp': '2025-11-27 10:17:44.253986', '_unique_id': '1b09ffa8933f43108ae8dd10a0df8578'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.254 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.256 12 INFO ceilometer.polling.manager [-] Polling pollster cpu in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.256 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/cpu volume: 19880000000 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': 'a852f794-f472-48dc-bdb5-f06416913d9c', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'cpu', 'counter_type': 'cumulative', 'counter_unit': 'ns', 'counter_volume': 19880000000, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'timestamp': '2025-11-27T10:17:44.256356', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'cpu_number': 1}, 'message_id': '50533ce2-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.357875375, 'message_signature': 'd78bac5df0f042b9dca6c0cd3d59db2769b1701f24952dc451576d85ec0252ff'}]}, 'timestamp': '2025-11-27 10:17:44.256857', '_unique_id': 'a55856fdb1754e44b44ddf8617535af7'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.257 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.259 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.write.requests in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 47 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.259 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.write.requests volume: 1 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '9511fa92-ef15-451f-a7b4-f630b5e1e023', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 47, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.259263', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5053ac86-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '9df34e341b436874afea333529ccc22a3b3e159a9dc32cd0a928cbc80c0519b6'}, {'source': 'openstack', 'counter_name': 'disk.device.write.requests', 'counter_type': 'cumulative', 'counter_unit': 'request', 'counter_volume': 1, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.259263', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '5053be10-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.383780861, 'message_signature': '4c4580295a37c4a9b649288f3d27f1405f3a24b783545d2f8d408b3ac7180c93'}]}, 'timestamp': '2025-11-27 10:17:44.260175', '_unique_id': '058a14a6b3fc44e7b270a76c7a8bf7e3'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 INFO ceilometer.polling.manager [-] Polling pollster network.incoming.packets.error in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.261 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/network.incoming.packets.error volume: 0 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '0376fdfb-b498-411d-baae-2a42601ad291', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'network.incoming.packets.error', 'counter_type': 'cumulative', 'counter_unit': 'packet', 'counter_volume': 0, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'instance-00000002-a02f7f2f-d2cf-4612-b18a-8be435257201-tapa2718872-3b', 'timestamp': '2025-11-27T10:17:44.261895', 'resource_metadata': {'display_name': 'test', 'name': 'tapa2718872-3b', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'mac': 'fa:16:3e:86:b9:b3', 'fref': None, 'parameters': {'interfaceid': None, 'bridge': None}, 'vnic_name': 'tapa2718872-3b'}, 'message_id': '50541022-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.332036321, 'message_signature': '2f0c20a87e84f88ce0ea5eced65ba00d1fea05a7a91911342443c4ba0a84cbc3'}]}, 'timestamp': '2025-11-27 10:17:44.262191', '_unique_id': '56670e08a7134a19a386a9ddefbb3598'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.262 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.263 12 INFO ceilometer.polling.manager [-] Polling pollster disk.device.allocation in the context of pollsters Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.263 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 DEBUG ceilometer.compute.pollsters [-] a02f7f2f-d2cf-4612-b18a-8be435257201/disk.device.allocation volume: 1073741824 _stats_to_sample /usr/lib/python3.9/site-packages/ceilometer/compute/pollsters/__init__.py:108 Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging [-] Could not send notification to notifications. Payload={'message_id': '97fcad5b-0c52-4962-9c11-7948ad5574a8', 'publisher_id': 'ceilometer.polling', 'event_type': 'telemetry.polling', 'priority': 'SAMPLE', 'payload': {'samples': [{'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vda', 'timestamp': '2025-11-27T10:17:44.263713', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vda'}, 'message_id': '5054571c-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': 'ba1682ef419693ba2dcf778bdf62d1c72029cb64c2e0703a8bfbe7167faeeb82'}, {'source': 'openstack', 'counter_name': 'disk.device.allocation', 'counter_type': 'gauge', 'counter_unit': 'B', 'counter_volume': 1073741824, 'user_id': '3911b9fb06a84c8eb08d5c917916dc45', 'user_name': None, 'project_id': '02a308de13ca4518beaddd402ba93936', 'project_name': None, 'resource_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201-vdb', 'timestamp': '2025-11-27T10:17:44.263713', 'resource_metadata': {'display_name': 'test', 'name': 'instance-00000002', 'instance_id': 'a02f7f2f-d2cf-4612-b18a-8be435257201', 'instance_type': 'm1.small', 'host': '2c5f75eca5573782b1273cbdbebced9c63eb09767c4b037e8b6c073d', 'instance_host': 'np0005537446.localdomain', 'flavor': {'id': 'ad8d4a49-18f7-4954-9703-7480b3ae8896', 'name': 'm1.small', 'vcpus': 1, 'ram': 512, 'disk': 1, 'ephemeral': 1, 'swap': 0}, 'status': 'active', 'state': 'running', 'task_state': '', 'image': {'id': '665842af-e631-4036-a0eb-09cf83c9a80f'}, 'image_ref': '665842af-e631-4036-a0eb-09cf83c9a80f', 'image_ref_url': None, 'architecture': 'x86_64', 'os_type': 'hvm', 'vcpus': 1, 'memory_mb': 512, 'disk_gb': 1, 'ephemeral_gb': 1, 'root_gb': 0, 'disk_name': 'vdb'}, 'message_id': '505461da-cb7a-11f0-8166-fa163e6bfee1', 'monotonic_time': 12894.361204924, 'message_signature': '2353f26cbb07414bb117a19575630306c23ea6218f9c0f40a844b5131192b9dc'}]}, 'timestamp': '2025-11-27 10:17:44.264263', '_unique_id': '2a0f0cd3905a4651b950620eedfa4bb8'}: kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 446, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging yield Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/utils/functional.py", line 312, in retry_over_time Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return fun(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 877, in _connection_factory Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self._connection = self._establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 812, in _establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging conn = self.transport.establish_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/transport/pyamqp.py", line 201, in establish_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging conn.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/connection.py", line 323, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.transport.connect() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 129, in connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self._connect(self.host, self.port, self.connect_timeout) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/amqp/transport.py", line 184, in _connect Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.sock.connect(sa) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging ConnectionRefusedError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging The above exception was the direct cause of the following exception: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging Traceback (most recent call last): Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/notify/messaging.py", line 78, in notify Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.transport._send_notification(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/transport.py", line 134, in _send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self._driver.send_notification(target, ctxt, message, version, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 694, in send_notification Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return self._send(target, ctxt, message, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 653, in _send Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging with self._get_connection(rpc_common.PURPOSE_SEND, retry) as conn: Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 605, in _get_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return rpc_common.ConnectionContext(self._connection_pool, Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/common.py", line 423, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.connection = connection_pool.get(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 98, in get Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return self.create(retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/pool.py", line 135, in create Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return self.connection_cls(self.conf, self.url, purpose, retry=retry) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 826, in __init__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.ensure_connection() Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/oslo_messaging/_drivers/impl_rabbit.py", line 957, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.connection.ensure_connection( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 381, in ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self._ensure_connection(*args, **kwargs) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 433, in _ensure_connection Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging return retry_over_time( Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib64/python3.9/contextlib.py", line 137, in __exit__ Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging self.gen.throw(typ, value, traceback) Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging File "/usr/lib/python3.9/site-packages/kombu/connection.py", line 450, in _reraise_as_library_errors Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging raise ConnectionError(str(exc)) from exc Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging kombu.exceptions.OperationalError: [Errno 111] Connection refused Nov 27 05:17:44 localhost ceilometer_agent_compute[240028]: 2025-11-27 10:17:44.264 12 ERROR oslo_messaging.notify.messaging Nov 27 05:17:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:17:44 Nov 27 05:17:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:17:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:17:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['vms', 'volumes', 'manila_data', 'images', 'manila_metadata', '.mgr', 'backups'] Nov 27 05:17:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:17:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 5.452610273590173e-07 of space, bias 1.0, pg target 0.00010850694444444444 quantized to 32 (current 32) Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:17:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.0024073274357900615 of space, bias 4.0, pg target 1.916232638888889 quantized to 16 (current 16) Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:17:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v730: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 47 KiB/s wr, 2 op/s Nov 27 05:17:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:17:45 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:17:45 localhost podman[332260]: 2025-11-27 10:17:45.992693858 +0000 UTC m=+0.086500294 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, org.label-schema.license=GPLv2, config_id=ovn_metadata_agent, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) Nov 27 05:17:46 localhost podman[332260]: 2025-11-27 10:17:46.002797149 +0000 UTC m=+0.096603615 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent) Nov 27 05:17:46 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:17:46 localhost podman[332261]: 2025-11-27 10:17:46.095341195 +0000 UTC m=+0.185794062 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_id=multipathd, container_name=multipathd, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.build-date=20251125) Nov 27 05:17:46 localhost podman[332261]: 2025-11-27 10:17:46.110995336 +0000 UTC m=+0.201448213 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, tcib_managed=true, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, container_name=multipathd, org.label-schema.schema-version=1.0) Nov 27 05:17:46 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:17:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:17:46.311 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:17:45Z, description=, device_id=b33b9f09-c017-4f44-bb21-0114094fbe1c, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=b1b23bd1-a63b-421c-a7c5-80179d45c85f, ip_allocation=immediate, mac_address=fa:16:3e:97:fe:16, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3909, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:17:45Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:17:46 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:17:46 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:17:46 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:17:46 localhost podman[332312]: 2025-11-27 10:17:46.623176854 +0000 UTC m=+0.058952644 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.733 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.905 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.905 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.906 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.906 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:17:46 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:17:46.915 265123 INFO neutron.agent.dhcp.agent [None req-13e6de95-d00e-4c66-87e5-9278625ca04f - - - - - -] DHCP configuration for ports {'b1b23bd1-a63b-421c-a7c5-80179d45c85f'} is completed#033[00m Nov 27 05:17:46 localhost nova_compute[284026]: 2025-11-27 10:17:46.990 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:46 localhost systemd[1]: tmp-crun.KmCkdY.mount: Deactivated successfully. Nov 27 05:17:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v731: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 47 KiB/s wr, 2 op/s Nov 27 05:17:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3b393b3f-0cd7-4dce-b7ff-34036516ff96_8bf15a37-734b-44c4-9c57-c95430a87cd2", "force": true, "format": "json"}]: dispatch Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96_8bf15a37-734b-44c4-9c57-c95430a87cd2, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96_8bf15a37-734b-44c4-9c57-c95430a87cd2, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:47 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume snapshot rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "snap_name": "3b393b3f-0cd7-4dce-b7ff-34036516ff96", "force": true, "format": "json"}]: dispatch Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] wrote 155 bytes to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.metadata_manager] Renamed b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta.tmp' to config b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056/.meta' Nov 27 05:17:47 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_snapshot_rm(force:True, format:json, prefix:fs subvolume snapshot rm, snap_name:3b393b3f-0cd7-4dce-b7ff-34036516ff96, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e276 e276: 6 total, 6 up, 6 in Nov 27 05:17:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e276 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.392 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.415 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.441 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.441 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.442 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._run_pending_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.443 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11145#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.631 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] There are 0 instances to clean _run_pending_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:11154#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.632 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:48 localhost nova_compute[284026]: 2025-11-27 10:17:48.662 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:17:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2085959973' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:17:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:17:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/2085959973' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:17:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v733: 177 pgs: 177 active+clean; 226 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 204 B/s rd, 47 KiB/s wr, 2 op/s Nov 27 05:17:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs clone status", "vol_name": "cephfs", "clone_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "format": "json"}]: dispatch Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_clone_status(clone_name:6edfae9d-e63a-4837-aaa6-0a145670e056, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_clone_status(clone_name:6edfae9d-e63a-4837-aaa6-0a145670e056, format:json, prefix:fs clone status, vol_name:cephfs) < "" Nov 27 05:17:50 localhost ceph-e83f3b0c-4090-52df-95d4-ad9be8516692-mgr-np0005537446-nfulyg[290373]: 2025-11-27T10:17:50.841+0000 7f54df954640 -1 mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '6edfae9d-e63a-4837-aaa6-0a145670e056' of type subvolume Nov 27 05:17:50 localhost ceph-mgr[290377]: mgr.server reply reply (95) Operation not supported operation 'clone-status' is not allowed on subvolume '6edfae9d-e63a-4837-aaa6-0a145670e056' of type subvolume Nov 27 05:17:50 localhost ceph-mgr[290377]: log_channel(audit) log [DBG] : from='client.25357 -' entity='client.openstack' cmd=[{"prefix": "fs subvolume rm", "vol_name": "cephfs", "sub_name": "6edfae9d-e63a-4837-aaa6-0a145670e056", "force": true, "format": "json"}]: dispatch Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Starting _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.operations.versions.subvolume_base] subvolume path 'b'/volumes/_nogroup/6edfae9d-e63a-4837-aaa6-0a145670e056'' moved to trashcan Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.fs.async_job] queuing job for volume 'cephfs' Nov 27 05:17:50 localhost ceph-mgr[290377]: [volumes INFO volumes.module] Finishing _cmd_fs_subvolume_rm(force:True, format:json, prefix:fs subvolume rm, sub_name:6edfae9d-e63a-4837-aaa6-0a145670e056, vol_name:cephfs) < "" Nov 27 05:17:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v734: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 409 B/s rd, 62 KiB/s wr, 4 op/s Nov 27 05:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:17:51 localhost ceph-osd[31760]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 9000.1 total, 600.0 interval#012Cumulative writes: 19K writes, 75K keys, 19K commit groups, 1.0 writes per commit group, ingest: 0.05 GB, 0.01 MB/s#012Cumulative WAL: 19K writes, 6551 syncs, 3.03 writes per sync, written: 0.05 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 10K writes, 37K keys, 10K commit groups, 1.0 writes per commit group, ingest: 20.32 MB, 0.03 MB/s#012Interval WAL: 10K writes, 4312 syncs, 2.53 writes per sync, written: 0.02 GB, 0.03 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 05:17:51 localhost nova_compute[284026]: 2025-11-27 10:17:51.992 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:53 localhost nova_compute[284026]: 2025-11-27 10:17:53.041 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e276 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e277 e277: 6 total, 6 up, 6 in Nov 27 05:17:53 localhost nova_compute[284026]: 2025-11-27 10:17:53.394 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v736: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 47 KiB/s wr, 3 op/s Nov 27 05:17:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e278 e278: 6 total, 6 up, 6 in Nov 27 05:17:54 localhost nova_compute[284026]: 2025-11-27 10:17:54.675 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._sync_power_states run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:54 localhost nova_compute[284026]: 2025-11-27 10:17:54.697 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Triggering sync for uuid a02f7f2f-d2cf-4612-b18a-8be435257201 _sync_power_states /usr/lib/python3.9/site-packages/nova/compute/manager.py:10268#033[00m Nov 27 05:17:54 localhost nova_compute[284026]: 2025-11-27 10:17:54.698 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "a02f7f2f-d2cf-4612-b18a-8be435257201" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:17:54 localhost nova_compute[284026]: 2025-11-27 10:17:54.698 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:17:54 localhost nova_compute[284026]: 2025-11-27 10:17:54.726 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "a02f7f2f-d2cf-4612-b18a-8be435257201" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.028s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:17:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v738: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 556 B/s rd, 77 KiB/s wr, 4 op/s Nov 27 05:17:55 localhost openstack_network_exporter[244641]: ERROR 10:17:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:17:55 localhost openstack_network_exporter[244641]: ERROR 10:17:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:17:55 localhost openstack_network_exporter[244641]: ERROR 10:17:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:17:55 localhost openstack_network_exporter[244641]: ERROR 10:17:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:17:55 localhost openstack_network_exporter[244641]: Nov 27 05:17:55 localhost openstack_network_exporter[244641]: ERROR 10:17:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:17:55 localhost openstack_network_exporter[244641]: Nov 27 05:17:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:17:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:17:55 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:17:55 localhost podman[332334]: 2025-11-27 10:17:55.991748184 +0000 UTC m=+0.082533747 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, container_name=ceilometer_agent_compute, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=edpm, io.buildah.version=1.41.3) Nov 27 05:17:56 localhost systemd[1]: tmp-crun.2x6Zg6.mount: Deactivated successfully. Nov 27 05:17:56 localhost podman[332336]: 2025-11-27 10:17:56.050355488 +0000 UTC m=+0.132776197 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, com.redhat.component=ubi9-minimal-container, container_name=openstack_network_exporter, distribution-scope=public, vendor=Red Hat, Inc., version=9.6, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, architecture=x86_64, io.openshift.tags=minimal rhel9, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., maintainer=Red Hat, Inc., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.buildah.version=1.33.7, release=1755695350, vcs-type=git, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, managed_by=edpm_ansible, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., config_id=edpm, name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers) Nov 27 05:17:56 localhost podman[332335]: 2025-11-27 10:17:56.02358082 +0000 UTC m=+0.110021697 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:17:56 localhost podman[332334]: 2025-11-27 10:17:56.081144666 +0000 UTC m=+0.171930239 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, org.label-schema.license=GPLv2, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, config_id=edpm, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:17:56 localhost podman[332336]: 2025-11-27 10:17:56.092941562 +0000 UTC m=+0.175362221 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, com.redhat.component=ubi9-minimal-container, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, io.buildah.version=1.33.7, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, architecture=x86_64, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, release=1755695350, distribution-scope=public, managed_by=edpm_ansible, name=ubi9-minimal, version=9.6, build-date=2025-08-20T13:12:41, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, maintainer=Red Hat, Inc., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-type=git, config_id=edpm, vendor=Red Hat, Inc.) Nov 27 05:17:56 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:17:56 localhost podman[332335]: 2025-11-27 10:17:56.107036071 +0000 UTC m=+0.193476988 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter) Nov 27 05:17:56 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:17:56 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:17:56 localhost ceph-osd[32722]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 9000.2 total, 600.0 interval#012Cumulative writes: 25K writes, 100K keys, 25K commit groups, 1.0 writes per commit group, ingest: 0.09 GB, 0.01 MB/s#012Cumulative WAL: 25K writes, 9048 syncs, 2.84 writes per sync, written: 0.09 GB, 0.01 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 14K writes, 53K keys, 14K commit groups, 1.0 writes per commit group, ingest: 53.80 MB, 0.09 MB/s#012Interval WAL: 14K writes, 5787 syncs, 2.44 writes per sync, written: 0.05 GB, 0.09 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent Nov 27 05:17:56 localhost systemd[1]: tmp-crun.62Yb8G.mount: Deactivated successfully. Nov 27 05:17:56 localhost nova_compute[284026]: 2025-11-27 10:17:56.994 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v739: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 511 B/s rd, 71 KiB/s wr, 4 op/s Nov 27 05:17:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e278 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:17:58 localhost nova_compute[284026]: 2025-11-27 10:17:58.446 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:58 localhost nova_compute[284026]: 2025-11-27 10:17:58.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._cleanup_incomplete_migrations run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:17:58 localhost nova_compute[284026]: 2025-11-27 10:17:58.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Cleaning up deleted instances with incomplete migration _cleanup_incomplete_migrations /usr/lib/python3.9/site-packages/nova/compute/manager.py:11183#033[00m Nov 27 05:17:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 e279: 6 total, 6 up, 6 in Nov 27 05:17:59 localhost ovn_controller[156436]: 2025-11-27T10:17:59Z|00351|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:17:59 localhost systemd[1]: tmp-crun.9TufgL.mount: Deactivated successfully. Nov 27 05:17:59 localhost nova_compute[284026]: 2025-11-27 10:17:59.177 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:17:59 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:17:59 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:17:59 localhost podman[332416]: 2025-11-27 10:17:59.180862622 +0000 UTC m=+0.059006407 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, maintainer=OpenStack Kubernetes Operator team, org.label-schema.schema-version=1.0, tcib_managed=true, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20251125) Nov 27 05:17:59 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:17:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v741: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 325 B/s rd, 31 KiB/s wr, 1 op/s Nov 27 05:18:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v742: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 255 B/s rd, 57 KiB/s wr, 2 op/s Nov 27 05:18:01 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:18:01.678 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:18:01Z, description=, device_id=8c66858f-378a-4896-a580-c9f03dc76fdc, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=7ff1369b-f910-4c97-bbf3-c33a52a14162, ip_allocation=immediate, mac_address=fa:16:3e:b9:42:78, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3943, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:18:01Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:18:01 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:18:01 localhost podman[332452]: 2025-11-27 10:18:01.902792759 +0000 UTC m=+0.063290411 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, tcib_managed=true) Nov 27 05:18:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:18:01 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:18:02 localhost nova_compute[284026]: 2025-11-27 10:18:02.042 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:02 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:18:02.236 265123 INFO neutron.agent.dhcp.agent [None req-bd60dd7f-4e18-497f-addb-8b6145d12c8c - - - - - -] DHCP configuration for ports {'7ff1369b-f910-4c97-bbf3-c33a52a14162'} is completed#033[00m Nov 27 05:18:02 localhost nova_compute[284026]: 2025-11-27 10:18:02.925 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:03 localhost nova_compute[284026]: 2025-11-27 10:18:03.448 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v743: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 212 B/s rd, 47 KiB/s wr, 2 op/s Nov 27 05:18:05 localhost nova_compute[284026]: 2025-11-27 10:18:05.409 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v744: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 27 KiB/s wr, 1 op/s Nov 27 05:18:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:18:05 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:18:06 localhost podman[332474]: 2025-11-27 10:18:06.011629633 +0000 UTC m=+0.094180161 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:18:06 localhost podman[332473]: 2025-11-27 10:18:06.055120691 +0000 UTC m=+0.141206944 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, container_name=ovn_controller, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_controller, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:18:06 localhost podman[332474]: 2025-11-27 10:18:06.076228728 +0000 UTC m=+0.158779246 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible) Nov 27 05:18:06 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:18:06 localhost podman[332473]: 2025-11-27 10:18:06.125000518 +0000 UTC m=+0.211086771 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, tcib_managed=true, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_controller, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}) Nov 27 05:18:06 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:18:07 localhost nova_compute[284026]: 2025-11-27 10:18:07.045 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v745: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 26 KiB/s wr, 1 op/s Nov 27 05:18:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:08 localhost nova_compute[284026]: 2025-11-27 10:18:08.483 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:08 localhost podman[242678]: time="2025-11-27T10:18:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:18:08 localhost podman[242678]: @ - - [27/Nov/2025:10:18:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:18:08 localhost podman[242678]: @ - - [27/Nov/2025:10:18:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19746 "" "Go-http-client/1.1" Nov 27 05:18:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v746: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 25 KiB/s wr, 1 op/s Nov 27 05:18:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v747: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 22 KiB/s wr, 0 op/s Nov 27 05:18:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Nov 27 05:18:11 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl.cc:1111] #012** DB Stats **#012Uptime(secs): 1200.0 total, 600.0 interval#012Cumulative writes: 4610 writes, 35K keys, 4610 commit groups, 1.0 writes per commit group, ingest: 0.06 GB, 0.05 MB/s#012Cumulative WAL: 4610 writes, 4610 syncs, 1.00 writes per sync, written: 0.06 GB, 0.05 MB/s#012Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent#012Interval writes: 2481 writes, 13K keys, 2481 commit groups, 1.0 writes per commit group, ingest: 18.89 MB, 0.03 MB/s#012Interval WAL: 2481 writes, 2481 syncs, 1.00 writes per sync, written: 0.02 GB, 0.03 MB/s#012Interval stall: 00:00:0.000 H:M:S, 0.0 percent#012#012** Compaction Stats [default] **#012Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 L0 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 146.0 0.28 0.11 16 0.018 0 0 0.0 0.0#012 L6 1/0 19.01 MB 0.0 0.3 0.0 0.3 0.3 0.0 0.0 6.9 136.1 126.1 2.27 0.81 15 0.151 202K 7721 0.0 0.0#012 Sum 1/0 19.01 MB 0.0 0.3 0.0 0.3 0.3 0.1 0.0 7.9 120.9 128.4 2.55 0.92 31 0.082 202K 7721 0.0 0.0#012 Int 0/0 0.00 KB 0.0 0.2 0.0 0.2 0.2 0.0 0.0 13.6 124.2 125.3 1.49 0.55 18 0.083 126K 4825 0.0 0.0#012#012** Compaction Stats [default] **#012Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB)#012---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------#012 Low 0/0 0.00 KB 0.0 0.3 0.0 0.3 0.3 0.0 0.0 0.0 136.1 126.1 2.27 0.81 15 0.151 202K 7721 0.0 0.0#012High 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 147.3 0.28 0.11 15 0.019 0 0 0.0 0.0#012User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.7 0.00 0.00 1 0.003 0 0 0.0 0.0#012#012Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0#012#012Uptime(secs): 1200.0 total, 600.0 interval#012Flush(GB): cumulative 0.041, interval 0.013#012AddFile(GB): cumulative 0.000, interval 0.000#012AddFile(Total Files): cumulative 0, interval 0#012AddFile(L0 Files): cumulative 0, interval 0#012AddFile(Keys): cumulative 0, interval 0#012Cumulative compaction: 0.32 GB write, 0.27 MB/s write, 0.30 GB read, 0.26 MB/s read, 2.6 seconds#012Interval compaction: 0.18 GB write, 0.31 MB/s write, 0.18 GB read, 0.31 MB/s read, 1.5 seconds#012Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count#012Block cache BinnedLRUCache@0x55fa1a6d5350#2 capacity: 304.00 MB usage: 24.49 MB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 0 last_secs: 0.000169 secs_since: 0#012Block cache entry stats(count,size,portion): DataBlock(1313,23.18 MB,7.62523%) FilterBlock(31,581.17 KB,0.186694%) IndexBlock(31,760.64 KB,0.244346%) Misc(1,0.00 KB,0%)#012#012** File Read Latency Histogram By Level [default] ** Nov 27 05:18:12 localhost nova_compute[284026]: 2025-11-27 10:18:12.047 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:13 localhost nova_compute[284026]: 2025-11-27 10:18:13.485 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v748: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:13 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:18:13.644 265123 INFO neutron.agent.dhcp.agent [-] Trigger reload_allocations for port admin_state_up=True, allowed_address_pairs=[], binding:host_id=, binding:profile=, binding:vif_details=, binding:vif_type=unbound, binding:vnic_type=normal, created_at=2025-11-27T10:18:12Z, description=, device_id=09f1f06a-d576-497c-956f-57456d68fe25, device_owner=network:router_gateway, dns_assignment=[], dns_domain=, dns_name=, extra_dhcp_opts=[], fixed_ips=[], id=d4605527-1d4f-4f8a-b82b-6cb55243b34e, ip_allocation=immediate, mac_address=fa:16:3e:7e:c3:04, name=, network=admin_state_up=True, availability_zone_hints=[], availability_zones=[], created_at=2025-11-27T08:32:23Z, description=, dns_domain=, id=38b54160-bf0a-4463-b656-cfe2a3341789, ipv4_address_scope=None, ipv6_address_scope=None, is_default=False, l2_adjacency=True, mtu=1350, name=public, port_security_enabled=True, project_id=02a308de13ca4518beaddd402ba93936, provider:network_type=flat, provider:physical_network=datacentre, provider:segmentation_id=None, qos_policy_id=None, revision_number=2, router:external=True, shared=False, standard_attr_id=29, status=ACTIVE, subnets=['a5bdef27-4a51-4d2a-96ef-9ef49863f717'], tags=[], tenant_id=02a308de13ca4518beaddd402ba93936, updated_at=2025-11-27T08:32:29Z, vlan_transparent=None, network_id=38b54160-bf0a-4463-b656-cfe2a3341789, port_security_enabled=False, project_id=, qos_network_policy_id=None, qos_policy_id=None, resource_request=None, revision_number=1, security_groups=[], standard_attr_id=3957, status=DOWN, tags=[], tenant_id=, updated_at=2025-11-27T10:18:12Z on network 38b54160-bf0a-4463-b656-cfe2a3341789#033[00m Nov 27 05:18:13 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 3 addresses Nov 27 05:18:13 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:18:13 localhost podman[332538]: 2025-11-27 10:18:13.869474533 +0000 UTC m=+0.061904014 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:18:13 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:18:13 localhost systemd[1]: tmp-crun.UWgbOH.mount: Deactivated successfully. Nov 27 05:18:14 localhost neutron_dhcp_agent[265119]: 2025-11-27 10:18:14.190 265123 INFO neutron.agent.dhcp.agent [None req-ae436bd1-108b-4eb3-9086-581f743a8b29 - - - - - -] DHCP configuration for ports {'d4605527-1d4f-4f8a-b82b-6cb55243b34e'} is completed#033[00m Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:14 localhost nova_compute[284026]: 2025-11-27 10:18:14.947 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v749: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:18:16 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:18:16 localhost podman[332576]: 2025-11-27 10:18:16.649675485 +0000 UTC m=+0.094185060 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, container_name=ovn_metadata_agent, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_managed=true, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:18:16 localhost podman[332576]: 2025-11-27 10:18:16.682108567 +0000 UTC m=+0.126618092 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=ovn_metadata_agent, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true) Nov 27 05:18:16 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:18:16 localhost podman[332577]: 2025-11-27 10:18:16.692825445 +0000 UTC m=+0.135704347 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, managed_by=edpm_ansible, config_id=multipathd, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, tcib_managed=true, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:18:16 localhost podman[332577]: 2025-11-27 10:18:16.775024132 +0000 UTC m=+0.217902984 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=multipathd, maintainer=OpenStack Kubernetes Operator team, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, container_name=multipathd, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, managed_by=edpm_ansible) Nov 27 05:18:16 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:18:17 localhost nova_compute[284026]: 2025-11-27 10:18:17.070 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:18:17 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:18:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:18:17 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:18:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:18:17 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 46bdcdf9-939b-4d4d-8531-92a4af5f2e7c (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:18:17 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 46bdcdf9-939b-4d4d-8531-92a4af5f2e7c (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:18:17 localhost ceph-mgr[290377]: [progress INFO root] Completed event 46bdcdf9-939b-4d4d-8531-92a4af5f2e7c (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:18:17 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:18:17 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:18:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v750: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:18 localhost nova_compute[284026]: 2025-11-27 10:18:17.998 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:18 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:18:18 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:18:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:18 localhost nova_compute[284026]: 2025-11-27 10:18:18.515 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v751: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:20 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:18:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:18:21 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:18:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v752: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:22 localhost nova_compute[284026]: 2025-11-27 10:18:22.104 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:22 localhost ovn_controller[156436]: 2025-11-27T10:18:22Z|00352|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:18:22 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 2 addresses Nov 27 05:18:22 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:18:22 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:18:22 localhost podman[332700]: 2025-11-27 10:18:22.374347212 +0000 UTC m=+0.069791676 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:18:22 localhost nova_compute[284026]: 2025-11-27 10:18:22.396 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v753: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:23 localhost nova_compute[284026]: 2025-11-27 10:18:23.516 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_write.cc:2098] [default] New memtable created with log file: #58. Immutable memtables: 0. Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.875589) [db/db_impl/db_impl_compaction_flush.cc:2832] Calling FlushMemTableToOutputFile with column family [default], flush slots available 1, compaction slots available 1, flush slots scheduled 1, compaction slots scheduled 0 Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:856] [default] [JOB 33] Flushing memtable with next log file: 58 Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238703875647, "job": 33, "event": "flush_started", "num_memtables": 1, "num_entries": 1001, "num_deletes": 259, "total_data_size": 1112736, "memory_usage": 1135416, "flush_reason": "Manual Compaction"} Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:885] [default] [JOB 33] Level-0 flush table #59: started Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238703882200, "cf_name": "default", "job": 33, "event": "table_file_creation", "file_number": 59, "file_size": 717025, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 35221, "largest_seqno": 36217, "table_properties": {"data_size": 712823, "index_size": 1865, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 1285, "raw_key_size": 9996, "raw_average_key_size": 19, "raw_value_size": 703945, "raw_average_value_size": 1393, "num_data_blocks": 82, "num_entries": 505, "num_filter_entries": 505, "num_deletions": 259, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764238649, "oldest_key_time": 1764238649, "file_creation_time": 1764238703, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 59, "seqno_to_time_mapping": "N/A"}} Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/flush_job.cc:1019] [default] [JOB 33] Flush lasted 6648 microseconds, and 3139 cpu microseconds. Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.882244) [db/flush_job.cc:967] [default] [JOB 33] Level-0 flush table #59: 717025 bytes OK Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.882265) [db/memtable_list.cc:519] [default] Level-0 commit table #59 started Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.885059) [db/memtable_list.cc:722] [default] Level-0 commit table #59: memtable #1 done Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.885079) EVENT_LOG_v1 {"time_micros": 1764238703885073, "job": 33, "event": "flush_finished", "output_compression": "NoCompression", "lsm_state": [1, 0, 0, 0, 0, 0, 1], "immutable_memtables": 0} Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.885100) [db/db_impl/db_impl_compaction_flush.cc:299] [default] Level summary: base level 6 level multiplier 10.00 max bytes base 268435456 files[1 0 0 0 0 0 1] max score 0.25 Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/db_impl/db_impl_files.cc:463] [JOB 33] Try to delete WAL files size 1107666, prev total WAL file size 1107990, number of live WAL files 2. Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000055.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.887934) [db/db_impl/db_impl_compaction_flush.cc:3165] [default] Manual compaction from level-0 to level-6 from '6C6F676D0034353230' seq:72057594037927935, type:22 .. '6C6F676D0034373731' seq:0, type:0; will stop at (end) Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 34] Compacting 1@0 + 1@6 files to L6, score -1.00 Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 33 Base level 0, inputs: [59(700KB)], [57(19MB)] Nov 27 05:18:23 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238703888049, "job": 34, "event": "compaction_started", "compaction_reason": "ManualCompaction", "files_L0": [59], "files_L6": [57], "score": -1, "input_data_size": 20654637, "oldest_snapshot_seqno": -1} Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: [db/compaction/compaction_job.cc:1588] [default] [JOB 34] Generated table #60: 14586 keys, 20529552 bytes, temperature: kUnknown Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238704024683, "cf_name": "default", "job": 34, "event": "table_file_creation", "file_number": 60, "file_size": 20529552, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 0, "largest_seqno": 0, "table_properties": {"data_size": 20445388, "index_size": 46623, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 36485, "raw_key_size": 390970, "raw_average_key_size": 26, "raw_value_size": 20196956, "raw_average_value_size": 1384, "num_data_blocks": 1743, "num_entries": 14586, "num_filter_entries": 14586, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1764237491, "oldest_key_time": 0, "file_creation_time": 1764238703, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a7e2bcde-7690-4ed3-a926-bff021ee997e", "db_session_id": "538JK4PSKJ851L30ETJB", "orig_file_number": 60, "seqno_to_time_mapping": "N/A"}} Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: [db/version_set.cc:4390] More existing levels in DB than needed. max_bytes_for_level_multiplier may not be guaranteed. Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.025000) [db/compaction/compaction_job.cc:1663] [default] [JOB 34] Compacted 1@0 + 1@6 files to L6 => 20529552 bytes Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.027054) [db/compaction/compaction_job.cc:865] [default] compacted to: base level 6 level multiplier 10.00 max bytes base 268435456 files[0 0 0 0 0 0 1] max score 0.00, MB/sec: 151.3 rd, 150.4 wr, level 6, files in(1, 1) out(1 +0 blob) MB in(0.7, 19.0 +0.0 blob) out(19.6 +0.0 blob), read-write-amplify(57.4) write-amplify(28.6) OK, records in: 15123, records dropped: 537 output_compression: NoCompression Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.027082) EVENT_LOG_v1 {"time_micros": 1764238704027069, "job": 34, "event": "compaction_finished", "compaction_time_micros": 136508, "compaction_time_cpu_micros": 54561, "output_level": 6, "num_output_files": 1, "total_output_size": 20529552, "num_input_records": 15123, "num_output_records": 14586, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [0, 0, 0, 0, 0, 0, 1]} Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000059.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238704027309, "job": 34, "event": "table_file_deletion", "file_number": 59} Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-np0005537446/store.db/000057.sst immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: EVENT_LOG_v1 {"time_micros": 1764238704030284, "job": 34, "event": "table_file_deletion", "file_number": 57} Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:23.887761) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.030376) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.030381) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.030384) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.030394) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ceph-mon[305520]: rocksdb: (Original Log Time 2025/11/27-10:18:24.030397) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Nov 27 05:18:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:24.879 162092 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=28, ssl=[], options={'arp_ns_explicit_output': 'true', 'fdb_removal_limit': '0', 'ignore_lsp_down': 'false', 'mac_binding_removal_limit': '0', 'mac_prefix': '86:30:76', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'f6:e7:8f:1d:ed:de'}, ipsec=False) old=SB_Global(nb_cfg=27) matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43#033[00m Nov 27 05:18:24 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:24.879 162092 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 3 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274#033[00m Nov 27 05:18:24 localhost nova_compute[284026]: 2025-11-27 10:18:24.880 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v754: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:25 localhost openstack_network_exporter[244641]: ERROR 10:18:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:18:25 localhost openstack_network_exporter[244641]: ERROR 10:18:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:18:25 localhost openstack_network_exporter[244641]: ERROR 10:18:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:18:25 localhost openstack_network_exporter[244641]: ERROR 10:18:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:18:25 localhost openstack_network_exporter[244641]: Nov 27 05:18:25 localhost openstack_network_exporter[244641]: ERROR 10:18:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:18:25 localhost openstack_network_exporter[244641]: Nov 27 05:18:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:18:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:18:26 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:18:27 localhost podman[332723]: 2025-11-27 10:18:27.008165066 +0000 UTC m=+0.094341066 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm) Nov 27 05:18:27 localhost podman[332723]: 2025-11-27 10:18:27.019918411 +0000 UTC m=+0.106094461 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:18:27 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:18:27 localhost podman[332722]: 2025-11-27 10:18:27.098658156 +0000 UTC m=+0.188556106 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.name=CentOS Stream 9 Base Image, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, config_id=edpm, org.label-schema.license=GPLv2) Nov 27 05:18:27 localhost nova_compute[284026]: 2025-11-27 10:18:27.108 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:27 localhost podman[332722]: 2025-11-27 10:18:27.142197026 +0000 UTC m=+0.232094996 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_managed=true, container_name=ceilometer_agent_compute, maintainer=OpenStack Kubernetes Operator team, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) Nov 27 05:18:27 localhost systemd[1]: tmp-crun.ZJORgg.mount: Deactivated successfully. Nov 27 05:18:27 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:18:27 localhost podman[332724]: 2025-11-27 10:18:27.145678019 +0000 UTC m=+0.227616485 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, container_name=openstack_network_exporter, maintainer=Red Hat, Inc., description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vendor=Red Hat, Inc., com.redhat.component=ubi9-minimal-container, io.buildah.version=1.33.7, managed_by=edpm_ansible, architecture=x86_64, release=1755695350, name=ubi9-minimal, io.openshift.expose-services=, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-type=git, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, config_id=edpm, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., version=9.6, distribution-scope=public, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., build-date=2025-08-20T13:12:41) Nov 27 05:18:27 localhost podman[332724]: 2025-11-27 10:18:27.228892915 +0000 UTC m=+0.310831421 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, config_id=edpm, io.openshift.tags=minimal rhel9, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., name=ubi9-minimal, url=https://catalog.redhat.com/en/search?searchType=containers, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, version=9.6, distribution-scope=public, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, container_name=openstack_network_exporter, release=1755695350, io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, maintainer=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.buildah.version=1.33.7, vcs-type=git, vendor=Red Hat, Inc., build-date=2025-08-20T13:12:41, managed_by=edpm_ansible) Nov 27 05:18:27 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:18:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v755: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:27 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:27.882 162092 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=927ada8e-9e05-4f20-b8c8-d7c6f26f86ee, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '28'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89#033[00m Nov 27 05:18:28 localhost ovn_controller[156436]: 2025-11-27T10:18:28Z|00353|binding|INFO|Releasing lport 60750ee5-275f-4987-8dcc-15b61f3cb86a from this chassis (sb_readonly=0) Nov 27 05:18:28 localhost nova_compute[284026]: 2025-11-27 10:18:28.102 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:28 localhost dnsmasq[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/addn_hosts - 1 addresses Nov 27 05:18:28 localhost podman[332800]: 2025-11-27 10:18:28.104483424 +0000 UTC m=+0.095334922 container kill 8627733b09c379e942bb48fc4f082dcd20dfc37b1a7cd4393c11d66b06e893f3 (image=quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified, name=neutron-dnsmasq-qdhcp-38b54160-bf0a-4463-b656-cfe2a3341789, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.build-date=20251125, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0) Nov 27 05:18:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/host Nov 27 05:18:28 localhost dnsmasq-dhcp[315688]: read /var/lib/neutron/dhcp/38b54160-bf0a-4463-b656-cfe2a3341789/opts Nov 27 05:18:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:28 localhost nova_compute[284026]: 2025-11-27 10:18:28.518 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v756: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v757: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:32 localhost nova_compute[284026]: 2025-11-27 10:18:32.142 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:32 localhost nova_compute[284026]: 2025-11-27 10:18:32.745 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rescued_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:32 localhost nova_compute[284026]: 2025-11-27 10:18:32.746 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:33 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:33 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v758: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:33 localhost nova_compute[284026]: 2025-11-27 10:18:33.521 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:34 localhost nova_compute[284026]: 2025-11-27 10:18:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_unconfirmed_resizes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:34 localhost nova_compute[284026]: 2025-11-27 10:18:34.731 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:34 localhost nova_compute[284026]: 2025-11-27 10:18:34.732 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._reclaim_queued_deletes run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:34 localhost nova_compute[284026]: 2025-11-27 10:18:34.732 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CONF.reclaim_instance_interval <= 0, skipping... _reclaim_queued_deletes /usr/lib/python3.9/site-packages/nova/compute/manager.py:10477#033[00m Nov 27 05:18:35 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v759: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:35 localhost nova_compute[284026]: 2025-11-27 10:18:35.733 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:18:36 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:18:37 localhost podman[332822]: 2025-11-27 10:18:37.061954253 +0000 UTC m=+0.156255778 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ovn_controller, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, config_id=ovn_controller, org.label-schema.vendor=CentOS, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3) Nov 27 05:18:37 localhost podman[332823]: 2025-11-27 10:18:37.035302167 +0000 UTC m=+0.125004728 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}, config_id=edpm, container_name=node_exporter) Nov 27 05:18:37 localhost podman[332823]: 2025-11-27 10:18:37.121059331 +0000 UTC m=+0.210761872 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:18:37 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:18:37 localhost podman[332822]: 2025-11-27 10:18:37.135978521 +0000 UTC m=+0.230280106 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, config_id=ovn_controller, container_name=ovn_controller, managed_by=edpm_ansible, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.schema-version=1.0, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, tcib_managed=true, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:18:37 localhost nova_compute[284026]: 2025-11-27 10:18:37.143 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:37 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:18:37 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v760: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:38 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.564 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.727 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.730 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager.update_available_resource run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.753 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.753 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.754 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.754 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Auditing locally available compute resources for np0005537446.localdomain (node: np0005537446.localdomain) update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:861#033[00m Nov 27 05:18:38 localhost nova_compute[284026]: 2025-11-27 10:18:38.754 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:18:38 localhost podman[242678]: time="2025-11-27T10:18:38Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:18:38 localhost podman[242678]: @ - - [27/Nov/2025:10:18:38 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:18:38 localhost podman[242678]: @ - - [27/Nov/2025:10:18:38 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19751 "" "Go-http-client/1.1" Nov 27 05:18:39 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:18:39 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/334952148' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.224 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.469s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.299 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.300 284030 DEBUG nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] skipping disk for instance-00000002 as it does not have a path _get_instance_disk_info_from_config /usr/lib/python3.9/site-packages/nova/virt/libvirt/driver.py:11231#033[00m Nov 27 05:18:39 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v761: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.534 284030 WARNING nova.virt.libvirt.driver [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] This host appears to have multiple sockets per NUMA node. The `socket` PCI NUMA affinity will not be supported.#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.536 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Hypervisor/Node resource view: name=np0005537446.localdomain free_ram=11024MB free_disk=41.83699035644531GB free_vcpus=7 pci_devices=[{"dev_id": "pci_0000_00_06_0", "address": "0000:00:06.0", "product_id": "1005", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1005", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_1", "address": "0000:00:01.1", "product_id": "7010", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7010", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_2", "address": "0000:00:01.2", "product_id": "7020", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7020", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_07_0", "address": "0000:00:07.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_0", "address": "0000:00:01.0", "product_id": "7000", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_03_0", "address": "0000:00:03.0", "product_id": "1000", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1000", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_01_3", "address": "0000:00:01.3", "product_id": "7113", "vendor_id": "8086", "numa_node": null, "label": "label_8086_7113", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_05_0", "address": "0000:00:05.0", "product_id": "1002", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1002", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_00_0", "address": "0000:00:00.0", "product_id": "1237", "vendor_id": "8086", "numa_node": null, "label": "label_8086_1237", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_04_0", "address": "0000:00:04.0", "product_id": "1001", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1001", "dev_type": "type-PCI"}, {"dev_id": "pci_0000_00_02_0", "address": "0000:00:02.0", "product_id": "1050", "vendor_id": "1af4", "numa_node": null, "label": "label_1af4_1050", "dev_type": "type-PCI"}] _report_hypervisor_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1034#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.536 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.537 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.619 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Instance a02f7f2f-d2cf-4612-b18a-8be435257201 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 512, 'VCPU': 1}}. _remove_deleted_instances_allocations /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1635#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.620 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Total usable vcpus: 8, total allocated vcpus: 1 _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1057#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.620 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Final resource view: name=np0005537446.localdomain phys_ram=15738MB used_ram=1024MB phys_disk=41GB used_disk=2GB total_vcpus=8 used_vcpus=1 pci_stats=[] _report_final_resource_view /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:1066#033[00m Nov 27 05:18:39 localhost nova_compute[284026]: 2025-11-27 10:18:39.673 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running cmd (subprocess): ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:384#033[00m Nov 27 05:18:40 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "df", "format": "json"} v 0) Nov 27 05:18:40 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.108:0/2689440496' entity='client.openstack' cmd={"prefix": "df", "format": "json"} : dispatch Nov 27 05:18:40 localhost nova_compute[284026]: 2025-11-27 10:18:40.123 284030 DEBUG oslo_concurrency.processutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] CMD "ceph df --format=json --id openstack --conf /etc/ceph/ceph.conf" returned: 0 in 0.449s execute /usr/lib/python3.9/site-packages/oslo_concurrency/processutils.py:422#033[00m Nov 27 05:18:40 localhost nova_compute[284026]: 2025-11-27 10:18:40.131 284030 DEBUG nova.compute.provider_tree [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed in ProviderTree for provider: 5764ad4c-cf89-4d5f-a185-92be71f7b67f update_inventory /usr/lib/python3.9/site-packages/nova/compute/provider_tree.py:180#033[00m Nov 27 05:18:40 localhost nova_compute[284026]: 2025-11-27 10:18:40.150 284030 DEBUG nova.scheduler.client.report [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Inventory has not changed for provider 5764ad4c-cf89-4d5f-a185-92be71f7b67f based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 16.0}, 'MEMORY_MB': {'total': 15738, 'reserved': 512, 'min_unit': 1, 'max_unit': 15738, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 41, 'reserved': 1, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} set_inventory_for_provider /usr/lib/python3.9/site-packages/nova/scheduler/client/report.py:940#033[00m Nov 27 05:18:40 localhost nova_compute[284026]: 2025-11-27 10:18:40.153 284030 DEBUG nova.compute.resource_tracker [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Compute_service record updated for np0005537446.localdomain:np0005537446.localdomain _update_available_resource /usr/lib/python3.9/site-packages/nova/compute/resource_tracker.py:995#033[00m Nov 27 05:18:40 localhost nova_compute[284026]: 2025-11-27 10:18:40.153 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.617s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:18:41 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v762: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:42 localhost nova_compute[284026]: 2025-11-27 10:18:42.146 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:43 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:43 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v763: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:43 localhost nova_compute[284026]: 2025-11-27 10:18:43.566 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:43.584 162092 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404#033[00m Nov 27 05:18:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:43.585 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409#033[00m Nov 27 05:18:43 localhost ovn_metadata_agent[162087]: 2025-11-27 10:18:43.586 162092 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423#033[00m Nov 27 05:18:44 localhost ceph-mgr[290377]: [balancer INFO root] Optimize plan auto_2025-11-27_10:18:44 Nov 27 05:18:44 localhost ceph-mgr[290377]: [balancer INFO root] Mode upmap, max misplaced 0.050000 Nov 27 05:18:44 localhost ceph-mgr[290377]: [balancer INFO root] do_upmap Nov 27 05:18:44 localhost ceph-mgr[290377]: [balancer INFO root] pools ['backups', 'manila_metadata', 'vms', 'images', 'volumes', '.mgr', 'manila_data'] Nov 27 05:18:44 localhost ceph-mgr[290377]: [balancer INFO root] prepared 0/10 changes Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:18:44 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] _maybe_adjust Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] TrashPurgeScheduleHandler: load_schedules Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool '.mgr' root_id -1 using 3.080724804578448e-05 of space, bias 1.0, pg target 0.006161449609156895 quantized to 1 (current 1) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'vms' root_id -1 using 0.0033250017448352874 of space, bias 1.0, pg target 0.6650003489670575 quantized to 32 (current 32) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'volumes' root_id -1 using 0.0014861089300670016 of space, bias 1.0, pg target 0.29672641637004465 quantized to 32 (current 32) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'images' root_id -1 using 0.004299383200725851 of space, bias 1.0, pg target 0.8584435124115949 quantized to 32 (current 32) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'backups' root_id -1 using 2.7263051367950866e-07 of space, bias 1.0, pg target 5.425347222222222e-05 quantized to 32 (current 32) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_data' root_id -1 using 0.0 of space, bias 1.0, pg target 0.0 quantized to 32 (current 32) Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] effective_target_ratio 0.0 0.0 0 45071990784 Nov 27 05:18:45 localhost ceph-mgr[290377]: [pg_autoscaler INFO root] Pool 'manila_metadata' root_id -1 using 0.002478211369346734 of space, bias 4.0, pg target 1.9726562500000002 quantized to 16 (current 16) Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] MirrorSnapshotScheduleHandler: load_schedules Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: vms, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: volumes, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: images, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: [rbd_support INFO root] load_schedules: backups, start_after= Nov 27 05:18:45 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v764: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:18:46 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:18:46 localhost podman[332915]: 2025-11-27 10:18:46.997226497 +0000 UTC m=+0.093568524 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:18:47 localhost podman[332915]: 2025-11-27 10:18:47.006956729 +0000 UTC m=+0.103298726 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, maintainer=OpenStack Kubernetes Operator team, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, config_id=ovn_metadata_agent, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.vendor=CentOS, tcib_managed=true) Nov 27 05:18:47 localhost podman[332916]: 2025-11-27 10:18:47.042119573 +0000 UTC m=+0.135238504 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=multipathd, managed_by=edpm_ansible) Nov 27 05:18:47 localhost podman[332916]: 2025-11-27 10:18:47.058095992 +0000 UTC m=+0.151214983 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team, org.label-schema.vendor=CentOS, container_name=multipathd, managed_by=edpm_ansible, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, config_id=multipathd, tcib_managed=true, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:18:47 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:18:47 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:18:47 localhost nova_compute[284026]: 2025-11-27 10:18:47.147 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:47 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v765: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.156 284030 DEBUG oslo_service.periodic_task [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.9/site-packages/oslo_service/periodic_task.py:210#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.156 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9858#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.157 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9862#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.619 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.667 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquiring lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.668 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Acquired lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.668 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Forcefully refreshing network info cache for instance _get_instance_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:2004#033[00m Nov 27 05:18:48 localhost nova_compute[284026]: 2025-11-27 10:18:48.668 284030 DEBUG nova.objects.instance [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Lazy-loading 'info_cache' on Instance uuid a02f7f2f-d2cf-4612-b18a-8be435257201 obj_load_attr /usr/lib/python3.9/site-packages/nova/objects/instance.py:1105#033[00m Nov 27 05:18:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"df", "format":"json"} v 0) Nov 27 05:18:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1223112081' entity='client.openstack' cmd={"prefix":"df", "format":"json"} : dispatch Nov 27 05:18:48 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} v 0) Nov 27 05:18:48 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='client.? 172.18.0.32:0/1223112081' entity='client.openstack' cmd={"prefix":"osd pool get-quota", "pool": "volumes", "format":"json"} : dispatch Nov 27 05:18:49 localhost nova_compute[284026]: 2025-11-27 10:18:49.171 284030 DEBUG nova.network.neutron [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updating instance_info_cache with network_info: [{"id": "a2718872-3b82-44ca-8d67-a0e516b2708b", "address": "fa:16:3e:86:b9:b3", "network": {"id": "3d77aad9-8375-4ff0-8227-0972f0af9ab9", "bridge": "br-int", "label": "private", "subnets": [{"cidr": "192.168.0.0/24", "dns": [], "gateway": {"address": "192.168.0.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.0.212", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "192.168.122.20", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.0.3"}}], "meta": {"injected": false, "tenant_id": "02a308de13ca4518beaddd402ba93936", "mtu": 1292, "physical_network": null, "tunneled": true}}, "type": "ovs", "details": {"port_filter": true, "connectivity": "l2", "datapath_type": "system", "bound_drivers": {"0": "ovn"}}, "devname": "tapa2718872-3b", "ovs_interfaceid": "a2718872-3b82-44ca-8d67-a0e516b2708b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] update_instance_cache_with_nw_info /usr/lib/python3.9/site-packages/nova/network/neutron.py:116#033[00m Nov 27 05:18:49 localhost nova_compute[284026]: 2025-11-27 10:18:49.187 284030 DEBUG oslo_concurrency.lockutils [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] Releasing lock "refresh_cache-a02f7f2f-d2cf-4612-b18a-8be435257201" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333#033[00m Nov 27 05:18:49 localhost nova_compute[284026]: 2025-11-27 10:18:49.188 284030 DEBUG nova.compute.manager [None req-3c7150e3-1f50-4571-98ba-4891835bcd6a - - - - - -] [instance: a02f7f2f-d2cf-4612-b18a-8be435257201] Updated the network info_cache for instance _heal_instance_info_cache /usr/lib/python3.9/site-packages/nova/compute/manager.py:9929#033[00m Nov 27 05:18:49 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v766: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:51 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v767: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:52 localhost nova_compute[284026]: 2025-11-27 10:18:52.149 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:53 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:53 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v768: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:53 localhost nova_compute[284026]: 2025-11-27 10:18:53.622 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:55 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v769: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:55 localhost openstack_network_exporter[244641]: ERROR 10:18:55 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:18:55 localhost openstack_network_exporter[244641]: ERROR 10:18:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:18:55 localhost openstack_network_exporter[244641]: ERROR 10:18:55 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:18:55 localhost openstack_network_exporter[244641]: ERROR 10:18:55 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:18:55 localhost openstack_network_exporter[244641]: Nov 27 05:18:55 localhost openstack_network_exporter[244641]: ERROR 10:18:55 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:18:55 localhost openstack_network_exporter[244641]: Nov 27 05:18:57 localhost nova_compute[284026]: 2025-11-27 10:18:57.151 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:57 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v770: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:18:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:18:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:18:57 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:18:58 localhost podman[332953]: 2025-11-27 10:18:58.003789445 +0000 UTC m=+0.092940888 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, container_name=ceilometer_agent_compute, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, config_id=edpm) Nov 27 05:18:58 localhost podman[332953]: 2025-11-27 10:18:58.043872802 +0000 UTC m=+0.133024235 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, config_id=edpm, container_name=ceilometer_agent_compute, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}, managed_by=edpm_ansible, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125) Nov 27 05:18:58 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:18:58 localhost podman[332954]: 2025-11-27 10:18:58.049258177 +0000 UTC m=+0.136326634 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:18:58 localhost podman[332955]: 2025-11-27 10:18:58.11194528 +0000 UTC m=+0.191308849 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, io.openshift.expose-services=, name=ubi9-minimal, vcs-type=git, io.buildah.version=1.33.7, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, architecture=x86_64, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, release=1755695350, vendor=Red Hat, Inc., maintainer=Red Hat, Inc., config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, distribution-scope=public, managed_by=edpm_ansible, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., com.redhat.component=ubi9-minimal-container, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, config_id=edpm, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., url=https://catalog.redhat.com/en/search?searchType=containers, build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, version=9.6) Nov 27 05:18:58 localhost podman[332954]: 2025-11-27 10:18:58.132971955 +0000 UTC m=+0.220040382 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi ) Nov 27 05:18:58 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:18:58 localhost podman[332955]: 2025-11-27 10:18:58.157147934 +0000 UTC m=+0.236511553 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, vendor=Red Hat, Inc., version=9.6, com.redhat.component=ubi9-minimal-container, maintainer=Red Hat, Inc., vcs-type=git, architecture=x86_64, io.buildah.version=1.33.7, url=https://catalog.redhat.com/en/search?searchType=containers, config_id=edpm, distribution-scope=public, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, build-date=2025-08-20T13:12:41, managed_by=edpm_ansible, release=1755695350, io.openshift.tags=minimal rhel9, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, name=ubi9-minimal, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, container_name=openstack_network_exporter) Nov 27 05:18:58 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:18:58 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:18:58 localhost nova_compute[284026]: 2025-11-27 10:18:58.653 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:18:59 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v771: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:00 localhost ovn_controller[156436]: 2025-11-27T10:19:00Z|00354|memory_trim|INFO|Detected inactivity (last active 30005 ms ago): trimming memory Nov 27 05:19:01 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v772: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:02 localhost nova_compute[284026]: 2025-11-27 10:19:02.153 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:03 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:03 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v773: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:03 localhost nova_compute[284026]: 2025-11-27 10:19:03.654 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:05 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v774: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:05 localhost sshd[333010]: main: sshd: ssh-rsa algorithm is disabled Nov 27 05:19:05 localhost systemd-logind[761]: New session 76 of user zuul. Nov 27 05:19:05 localhost systemd[1]: Started Session 76 of User zuul. Nov 27 05:19:06 localhost python3[333032]: ansible-ansible.legacy.command Invoked with _raw_params=subscription-manager unregister#012 _uses_shell=True zuul_log_id=fa163ec2-ffbe-e966-c683-00000000000c-1-overcloudnovacompute2 zuul_ansible_split_streams=False warn=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Nov 27 05:19:07 localhost nova_compute[284026]: 2025-11-27 10:19:07.182 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:07 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v775: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133. Nov 27 05:19:07 localhost systemd[1]: Started /usr/bin/podman healthcheck run d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a. Nov 27 05:19:08 localhost podman[333035]: 2025-11-27 10:19:08.002099101 +0000 UTC m=+0.086839573 container health_status c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, health_status=healthy, tcib_managed=true, org.label-schema.build-date=20251125, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_controller, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.license=GPLv2, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, org.label-schema.schema-version=1.0, config_id=ovn_controller) Nov 27 05:19:08 localhost systemd[1]: tmp-crun.qZbGoq.mount: Deactivated successfully. Nov 27 05:19:08 localhost podman[333036]: 2025-11-27 10:19:08.057801468 +0000 UTC m=+0.138839830 container health_status d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, health_status=healthy, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:19:08 localhost podman[333036]: 2025-11-27 10:19:08.067986872 +0000 UTC m=+0.149025214 container exec_died d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a (image=quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c, name=node_exporter, config_id=edpm, container_name=node_exporter, maintainer=The Prometheus Authors , managed_by=edpm_ansible, config_data={'image': 'quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9100:9100'], 'command': ['--web.disable-exporter-metrics', '--collector.systemd', '--collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\\.service', '--no-collector.dmi', '--no-collector.entropy', '--no-collector.thermal_zone', '--no-collector.time', '--no-collector.timex', '--no-collector.uname', '--no-collector.stat', '--no-collector.hwmon', '--no-collector.os', '--no-collector.selinux', '--no-collector.textfile', '--no-collector.powersupplyclass', '--no-collector.pressure', '--no-collector.rapl'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck node_exporter', 'mount': '/var/lib/openstack/healthchecks/node_exporter'}, 'volumes': ['/var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:rw', '/var/lib/openstack/healthchecks/node_exporter:/openstack:ro,z']}) Nov 27 05:19:08 localhost systemd[1]: d4558d785cab5d256e9389dd5e65962c626e5413108be410b427f0c8422f7d2a.service: Deactivated successfully. Nov 27 05:19:08 localhost podman[333035]: 2025-11-27 10:19:08.123403901 +0000 UTC m=+0.208144413 container exec_died c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133 (image=quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified, name=ovn_controller, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, config_data={'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_controller', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/lib/modules:/lib/modules:ro', '/run:/run', '/var/lib/openvswitch/ovn:/run/ovn:shared,z', '/var/lib/kolla/config_files/ovn_controller.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/openstack/cacerts/ovn/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_controller:/openstack:ro,z']}, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, container_name=ovn_controller, config_id=ovn_controller, org.label-schema.license=GPLv2) Nov 27 05:19:08 localhost systemd[1]: c6079102a18d380ee388403031e8eccce3d21dd31e13e2249aa39ab94a17c133.service: Deactivated successfully. Nov 27 05:19:08 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:08 localhost nova_compute[284026]: 2025-11-27 10:19:08.685 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:08 localhost podman[242678]: time="2025-11-27T10:19:08Z" level=info msg="List containers: received `last` parameter - overwriting `limit`" Nov 27 05:19:08 localhost podman[242678]: @ - - [27/Nov/2025:10:19:08 +0000] "GET /v4.9.3/libpod/containers/json?all=true&external=false&last=0&namespace=false&size=false&sync=false HTTP/1.1" 200 157512 "" "Go-http-client/1.1" Nov 27 05:19:08 localhost podman[242678]: @ - - [27/Nov/2025:10:19:08 +0000] "GET /v4.9.3/libpod/containers/stats?all=false&interval=1&stream=false HTTP/1.1" 200 19745 "" "Go-http-client/1.1" Nov 27 05:19:09 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v776: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:10 localhost systemd[1]: session-76.scope: Deactivated successfully. Nov 27 05:19:10 localhost systemd-logind[761]: Session 76 logged out. Waiting for processes to exit. Nov 27 05:19:10 localhost systemd-logind[761]: Removed session 76. Nov 27 05:19:11 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v777: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:12 localhost nova_compute[284026]: 2025-11-27 10:19:12.233 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:13 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:13 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v778: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail Nov 27 05:19:13 localhost nova_compute[284026]: 2025-11-27 10:19:13.723 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', )] Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [('cephfs', ), ('cephfs', )] Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] disconnecting from cephfs 'cephfs' Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] scanning for idle connections.. Nov 27 05:19:14 localhost ceph-mgr[290377]: [volumes INFO mgr_util] cleaning up connections: [] Nov 27 05:19:15 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v779: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 341 B/s wr, 0 op/s Nov 27 05:19:17 localhost nova_compute[284026]: 2025-11-27 10:19:17.272 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:17 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v780: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s wr, 0 op/s Nov 27 05:19:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51. Nov 27 05:19:17 localhost systemd[1]: Started /usr/bin/podman healthcheck run d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d. Nov 27 05:19:17 localhost podman[333099]: 2025-11-27 10:19:17.855336864 +0000 UTC m=+0.076287661 container health_status d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, health_status=healthy, org.label-schema.vendor=CentOS, managed_by=edpm_ansible, tcib_managed=true, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629) Nov 27 05:19:17 localhost podman[333099]: 2025-11-27 10:19:17.86897035 +0000 UTC m=+0.089921187 container exec_died d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d (image=quay.io/podified-antelope-centos9/openstack-multipathd:current-podified, name=multipathd, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, managed_by=edpm_ansible, org.label-schema.vendor=CentOS, container_name=multipathd, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, tcib_managed=true, config_data={'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/multipathd', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-multipathd:current-podified', 'net': 'host', 'privileged': True, 'restart': 'always', 'volumes': ['/etc/hosts:/etc/hosts:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '/dev/log:/dev/log', '/var/lib/kolla/config_files/multipathd.json:/var/lib/kolla/config_files/config.json:ro', '/dev:/dev', '/run/udev:/run/udev', '/sys:/sys', '/lib/modules:/lib/modules:ro', '/etc/iscsi:/etc/iscsi:ro', '/var/lib/iscsi:/var/lib/iscsi', '/etc/multipath:/etc/multipath:z', '/etc/multipath.conf:/etc/multipath.conf:ro', '/var/lib/openstack/healthchecks/multipathd:/openstack:ro,z']}, config_id=multipathd, org.label-schema.license=GPLv2, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.build-date=20251125) Nov 27 05:19:17 localhost systemd[1]: d3bf45c648d169722f0c40bfa75e3329fd837ab738feef834425e7f71a0a806d.service: Deactivated successfully. Nov 27 05:19:17 localhost podman[333097]: 2025-11-27 10:19:17.922758585 +0000 UTC m=+0.143364162 container health_status 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, health_status=healthy, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, container_name=ovn_metadata_agent, managed_by=edpm_ansible, org.label-schema.build-date=20251125, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, config_id=ovn_metadata_agent, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, tcib_managed=true, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, maintainer=OpenStack Kubernetes Operator team) Nov 27 05:19:17 localhost podman[333097]: 2025-11-27 10:19:17.956122561 +0000 UTC m=+0.176728168 container exec_died 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51 (image=quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified, name=ovn_metadata_agent, config_id=ovn_metadata_agent, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, tcib_managed=true, org.label-schema.schema-version=1.0, config_data={'cgroupns': 'host', 'depends_on': ['openvswitch.service'], 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'EDPM_CONFIG_HASH': 'df122b180261157f1de1391083b3d8abac306e2f12893ac7b9291feafc874311'}, 'healthcheck': {'mount': '/var/lib/openstack/healthchecks/ovn_metadata_agent', 'test': '/openstack/healthcheck'}, 'image': 'quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified', 'net': 'host', 'pid': 'host', 'privileged': True, 'restart': 'always', 'user': 'root', 'volumes': ['/run/openvswitch:/run/openvswitch:z', '/var/lib/config-data/ansible-generated/neutron-ovn-metadata-agent:/etc/neutron.conf.d:z', '/run/netns:/run/netns:shared', '/var/lib/kolla/config_files/ovn_metadata_agent.json:/var/lib/kolla/config_files/config.json:ro', '/var/lib/neutron:/var/lib/neutron:shared,z', '/var/lib/neutron/ovn_metadata_haproxy_wrapper:/usr/local/bin/haproxy:ro', '/var/lib/neutron/kill_scripts:/etc/neutron/kill_scripts:ro', '/var/lib/openstack/cacerts/neutron-metadata/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/var/lib/openstack/healthchecks/ovn_metadata_agent:/openstack:ro,z']}, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, container_name=ovn_metadata_agent, maintainer=OpenStack Kubernetes Operator team, managed_by=edpm_ansible, org.label-schema.name=CentOS Stream 9 Base Image) Nov 27 05:19:17 localhost systemd[1]: 1faf0be90bf4394ebc3c0653709e5aa58f7c29ac17006555fa84135a1e54fa51.service: Deactivated successfully. Nov 27 05:19:18 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:18 localhost nova_compute[284026]: 2025-11-27 10:19:18.763 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:19 localhost podman[333262]: Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.407461157 +0000 UTC m=+0.088829097 container create 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_CLEAN=True, ceph=True, architecture=x86_64, com.redhat.component=rhceph-container, version=7, distribution-scope=public, CEPH_POINT_RELEASE=, vcs-type=git, GIT_BRANCH=main, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, name=rhceph, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, maintainer=Guillaume Abrioux , release=553, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, description=Red Hat Ceph Storage 7, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0) Nov 27 05:19:19 localhost systemd[1]: Started libpod-conmon-3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb.scope. Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.370934166 +0000 UTC m=+0.052302166 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 05:19:19 localhost systemd[1]: Started libcrun container. Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.501287767 +0000 UTC m=+0.182655697 container init 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, build-date=2025-09-24T08:57:55, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., name=rhceph, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, description=Red Hat Ceph Storage 7, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.tags=rhceph ceph, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.buildah.version=1.33.12, maintainer=Guillaume Abrioux , version=7, GIT_CLEAN=True, vendor=Red Hat, Inc., architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, CEPH_POINT_RELEASE=, RELEASE=main, io.openshift.expose-services=, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7) Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.522092536 +0000 UTC m=+0.203460466 container start 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, maintainer=Guillaume Abrioux , architecture=x86_64, name=rhceph, io.openshift.expose-services=, description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-type=git, GIT_BRANCH=main, io.openshift.tags=rhceph ceph, GIT_CLEAN=True, RELEASE=main, distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, release=553, vendor=Red Hat, Inc., CEPH_POINT_RELEASE=, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_REPO=https://github.com/ceph/ceph-container.git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.buildah.version=1.33.12, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.k8s.description=Red Hat Ceph Storage 7, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, ceph=True) Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.522848056 +0000 UTC m=+0.204215986 container attach 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, description=Red Hat Ceph Storage 7, release=553, CEPH_POINT_RELEASE=, vendor=Red Hat, Inc., io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, distribution-scope=public, GIT_BRANCH=main, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.33.12, name=rhceph, io.openshift.tags=rhceph ceph, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, GIT_CLEAN=True, ceph=True, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.expose-services=, version=7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.k8s.description=Red Hat Ceph Storage 7, build-date=2025-09-24T08:57:55) Nov 27 05:19:19 localhost admiring_williamson[333276]: 167 167 Nov 27 05:19:19 localhost systemd[1]: libpod-3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb.scope: Deactivated successfully. Nov 27 05:19:19 localhost podman[333262]: 2025-11-27 10:19:19.531884189 +0000 UTC m=+0.213252189 container died 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, CEPH_POINT_RELEASE=, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, distribution-scope=public, GIT_CLEAN=True, io.buildah.version=1.33.12, architecture=x86_64, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., ceph=True, vendor=Red Hat, Inc., vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, io.k8s.description=Red Hat Ceph Storage 7, GIT_BRANCH=main, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, version=7, name=rhceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, description=Red Hat Ceph Storage 7, vcs-type=git, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, RELEASE=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, build-date=2025-09-24T08:57:55, release=553, com.redhat.license_terms=https://www.redhat.com/agreements, com.redhat.component=rhceph-container) Nov 27 05:19:19 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v781: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s wr, 0 op/s Nov 27 05:19:19 localhost podman[333281]: 2025-11-27 10:19:19.644853673 +0000 UTC m=+0.098587819 container remove 3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=admiring_williamson, ceph=True, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git, GIT_BRANCH=main, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, name=rhceph, com.redhat.license_terms=https://www.redhat.com/agreements, distribution-scope=public, maintainer=Guillaume Abrioux , architecture=x86_64, com.redhat.component=rhceph-container, CEPH_POINT_RELEASE=, io.openshift.expose-services=, io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, GIT_CLEAN=True, io.openshift.tags=rhceph ceph, vendor=Red Hat, Inc., RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, build-date=2025-09-24T08:57:55, description=Red Hat Ceph Storage 7, release=553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image.) Nov 27 05:19:19 localhost systemd[1]: libpod-conmon-3d9e6a7dea3c85a556674d516aaef60ab71f537a3231b4c06e027098aa53ecdb.scope: Deactivated successfully. Nov 27 05:19:19 localhost podman[333302]: Nov 27 05:19:19 localhost podman[333302]: 2025-11-27 10:19:19.894702636 +0000 UTC m=+0.086556076 container create bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, GIT_CLEAN=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, com.redhat.component=rhceph-container, maintainer=Guillaume Abrioux , io.openshift.expose-services=, io.buildah.version=1.33.12, RELEASE=main, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, vcs-type=git, version=7, release=553, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, name=rhceph, description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_BRANCH=main, architecture=x86_64, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., CEPH_POINT_RELEASE=, build-date=2025-09-24T08:57:55, ceph=True) Nov 27 05:19:19 localhost systemd[1]: Started libpod-conmon-bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3.scope. Nov 27 05:19:19 localhost podman[333302]: 2025-11-27 10:19:19.857419464 +0000 UTC m=+0.049272944 image pull registry.redhat.io/rhceph/rhceph-7-rhel9:latest Nov 27 05:19:19 localhost systemd[1]: Started libcrun container. Nov 27 05:19:19 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/36b6a41a592647784a89659495d68b1d88c1b6f420975e9f8bd016a4f291d02f/merged/rootfs supports timestamps until 2038 (0x7fffffff) Nov 27 05:19:19 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/36b6a41a592647784a89659495d68b1d88c1b6f420975e9f8bd016a4f291d02f/merged/var/log/ceph supports timestamps until 2038 (0x7fffffff) Nov 27 05:19:19 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/36b6a41a592647784a89659495d68b1d88c1b6f420975e9f8bd016a4f291d02f/merged/etc/ceph/ceph.conf supports timestamps until 2038 (0x7fffffff) Nov 27 05:19:19 localhost kernel: xfs filesystem being remounted at /var/lib/containers/storage/overlay/36b6a41a592647784a89659495d68b1d88c1b6f420975e9f8bd016a4f291d02f/merged/var/lib/ceph/crash supports timestamps until 2038 (0x7fffffff) Nov 27 05:19:19 localhost podman[333302]: 2025-11-27 10:19:19.975409913 +0000 UTC m=+0.167263353 container init bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, GIT_CLEAN=True, GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, com.redhat.component=rhceph-container, description=Red Hat Ceph Storage 7, distribution-scope=public, RELEASE=main, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, vcs-type=git, io.openshift.expose-services=, io.openshift.tags=rhceph ceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, io.k8s.description=Red Hat Ceph Storage 7, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., vendor=Red Hat, Inc., io.buildah.version=1.33.12, name=rhceph, release=553, build-date=2025-09-24T08:57:55, ceph=True, version=7, architecture=x86_64) Nov 27 05:19:19 localhost podman[333302]: 2025-11-27 10:19:19.986712267 +0000 UTC m=+0.178565707 container start bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., version=7, ceph=True, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, build-date=2025-09-24T08:57:55, vendor=Red Hat, Inc., io.openshift.tags=rhceph ceph, GIT_CLEAN=True, distribution-scope=public, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, maintainer=Guillaume Abrioux , GIT_REPO=https://github.com/ceph/ceph-container.git, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, name=rhceph, description=Red Hat Ceph Storage 7, architecture=x86_64, io.openshift.expose-services=, com.redhat.component=rhceph-container, release=553, RELEASE=main, GIT_BRANCH=main, com.redhat.license_terms=https://www.redhat.com/agreements, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, io.k8s.description=Red Hat Ceph Storage 7, vcs-type=git) Nov 27 05:19:19 localhost podman[333302]: 2025-11-27 10:19:19.988422223 +0000 UTC m=+0.180275733 container attach bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, maintainer=Guillaume Abrioux , GIT_BRANCH=main, CEPH_POINT_RELEASE=, version=7, build-date=2025-09-24T08:57:55, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, GIT_REPO=https://github.com/ceph/ceph-container.git, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, io.openshift.expose-services=, vcs-type=git, release=553, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, name=rhceph, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., io.buildah.version=1.33.12, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, ceph=True, distribution-scope=public, io.k8s.description=Red Hat Ceph Storage 7, vendor=Red Hat, Inc., GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, com.redhat.component=rhceph-container, architecture=x86_64) Nov 27 05:19:20 localhost systemd[1]: tmp-crun.P62LlT.mount: Deactivated successfully. Nov 27 05:19:20 localhost systemd[1]: var-lib-containers-storage-overlay-755116d0db8ff34921112b7e6b085f8f975a827ec0157ef35500acee9f843766-merged.mount: Deactivated successfully. Nov 27 05:19:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain.devices.0}] v 0) Nov 27 05:19:20 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537444.localdomain}] v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain.devices.0}] v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537445.localdomain}] v 0) Nov 27 05:19:21 localhost busy_williamson[333318]: [ Nov 27 05:19:21 localhost busy_williamson[333318]: { Nov 27 05:19:21 localhost busy_williamson[333318]: "available": false, Nov 27 05:19:21 localhost busy_williamson[333318]: "ceph_device": false, Nov 27 05:19:21 localhost busy_williamson[333318]: "device_id": "QEMU_DVD-ROM_QM00001", Nov 27 05:19:21 localhost busy_williamson[333318]: "lsm_data": {}, Nov 27 05:19:21 localhost busy_williamson[333318]: "lvs": [], Nov 27 05:19:21 localhost busy_williamson[333318]: "path": "/dev/sr0", Nov 27 05:19:21 localhost busy_williamson[333318]: "rejected_reasons": [ Nov 27 05:19:21 localhost busy_williamson[333318]: "Has a FileSystem", Nov 27 05:19:21 localhost busy_williamson[333318]: "Insufficient space (<5GB)" Nov 27 05:19:21 localhost busy_williamson[333318]: ], Nov 27 05:19:21 localhost busy_williamson[333318]: "sys_api": { Nov 27 05:19:21 localhost busy_williamson[333318]: "actuators": null, Nov 27 05:19:21 localhost busy_williamson[333318]: "device_nodes": "sr0", Nov 27 05:19:21 localhost busy_williamson[333318]: "human_readable_size": "482.00 KB", Nov 27 05:19:21 localhost busy_williamson[333318]: "id_bus": "ata", Nov 27 05:19:21 localhost busy_williamson[333318]: "model": "QEMU DVD-ROM", Nov 27 05:19:21 localhost busy_williamson[333318]: "nr_requests": "2", Nov 27 05:19:21 localhost busy_williamson[333318]: "partitions": {}, Nov 27 05:19:21 localhost busy_williamson[333318]: "path": "/dev/sr0", Nov 27 05:19:21 localhost busy_williamson[333318]: "removable": "1", Nov 27 05:19:21 localhost busy_williamson[333318]: "rev": "2.5+", Nov 27 05:19:21 localhost busy_williamson[333318]: "ro": "0", Nov 27 05:19:21 localhost busy_williamson[333318]: "rotational": "1", Nov 27 05:19:21 localhost busy_williamson[333318]: "sas_address": "", Nov 27 05:19:21 localhost busy_williamson[333318]: "sas_device_handle": "", Nov 27 05:19:21 localhost busy_williamson[333318]: "scheduler_mode": "mq-deadline", Nov 27 05:19:21 localhost busy_williamson[333318]: "sectors": 0, Nov 27 05:19:21 localhost busy_williamson[333318]: "sectorsize": "2048", Nov 27 05:19:21 localhost busy_williamson[333318]: "size": 493568.0, Nov 27 05:19:21 localhost busy_williamson[333318]: "support_discard": "0", Nov 27 05:19:21 localhost busy_williamson[333318]: "type": "disk", Nov 27 05:19:21 localhost busy_williamson[333318]: "vendor": "QEMU" Nov 27 05:19:21 localhost busy_williamson[333318]: } Nov 27 05:19:21 localhost busy_williamson[333318]: } Nov 27 05:19:21 localhost busy_williamson[333318]: ] Nov 27 05:19:21 localhost systemd[1]: libpod-bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3.scope: Deactivated successfully. Nov 27 05:19:21 localhost podman[333302]: 2025-11-27 10:19:21.131915079 +0000 UTC m=+1.323768509 container died bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, architecture=x86_64, name=rhceph, ceph=True, GIT_REPO=https://github.com/ceph/ceph-container.git, maintainer=Guillaume Abrioux , url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, GIT_BRANCH=main, vcs-type=git, GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, io.openshift.tags=rhceph ceph, description=Red Hat Ceph Storage 7, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, release=553, io.k8s.description=Red Hat Ceph Storage 7, io.openshift.expose-services=, com.redhat.component=rhceph-container, vendor=Red Hat, Inc., build-date=2025-09-24T08:57:55, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, GIT_CLEAN=True, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., distribution-scope=public, com.redhat.license_terms=https://www.redhat.com/agreements, CEPH_POINT_RELEASE=, io.buildah.version=1.33.12, version=7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, RELEASE=main) Nov 27 05:19:21 localhost systemd[1]: libpod-bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3.scope: Consumed 1.161s CPU time. Nov 27 05:19:21 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:21 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:21 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:21 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:21 localhost systemd[1]: var-lib-containers-storage-overlay-36b6a41a592647784a89659495d68b1d88c1b6f420975e9f8bd016a4f291d02f-merged.mount: Deactivated successfully. Nov 27 05:19:21 localhost podman[335375]: 2025-11-27 10:19:21.233396636 +0000 UTC m=+0.093732989 container remove bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3 (image=registry.redhat.io/rhceph/rhceph-7-rhel9:latest, name=busy_williamson, io.openshift.expose-services=, GIT_CLEAN=True, vcs-ref=cba612d428f1498c8ae5570dd75a670ccf94c03d, description=Red Hat Ceph Storage 7, ceph=True, name=rhceph, CEPH_POINT_RELEASE=, io.k8s.description=Red Hat Ceph Storage 7, org.opencontainers.image.revision=0c20ee48321f5d64135f6208d1332c0b032df6c3, GIT_BRANCH=main, distribution-scope=public, io.k8s.display-name=Red Hat Ceph Storage 7 on RHEL 9, summary=Provides the latest Red Hat Ceph Storage 7 on RHEL 9 in a fully featured and supported base image., GIT_COMMIT=12717c0777377369ea674892da98b0d85250f5b0, com.redhat.license_terms=https://www.redhat.com/agreements, RELEASE=main, GIT_REPO=https://github.com/ceph/ceph-container.git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/rhceph/images/7-553, architecture=x86_64, maintainer=Guillaume Abrioux , com.redhat.component=rhceph-container, vendor=Red Hat, Inc., version=7, io.openshift.tags=rhceph ceph, vcs-type=git, io.buildah.version=1.33.12, build-date=2025-09-24T08:57:55, release=553) Nov 27 05:19:21 localhost systemd[1]: libpod-conmon-bb031d820d70dcea4184a4b1a7a91665a10dc2225f65eaf6681622ada86dcbe3.scope: Deactivated successfully. Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain.devices.0}] v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/host.np0005537446.localdomain}] v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "config generate-minimal-conf"} v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "config generate-minimal-conf"} : dispatch Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "auth get", "entity": "client.admin"} v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: log_channel(audit) log [INF] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/cephadm/osd_remove_queue}] v 0) Nov 27 05:19:21 localhost ceph-mgr[290377]: [progress INFO root] update: starting ev 5219c8aa-662b-439e-981c-4d62764442e4 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:19:21 localhost ceph-mgr[290377]: [progress INFO root] complete: finished ev 5219c8aa-662b-439e-981c-4d62764442e4 (Updating node-proxy deployment (+3 -> 3)) Nov 27 05:19:21 localhost ceph-mgr[290377]: [progress INFO root] Completed event 5219c8aa-662b-439e-981c-4d62764442e4 (Updating node-proxy deployment (+3 -> 3)) in 0 seconds Nov 27 05:19:21 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command({"prefix": "osd tree", "states": ["destroyed"], "format": "json"} v 0) Nov 27 05:19:21 localhost ceph-mon[305520]: log_channel(audit) log [DBG] : from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch Nov 27 05:19:21 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v782: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s wr, 0 op/s Nov 27 05:19:22 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:22 localhost ceph-mon[305520]: from='mgr.34541 172.18.0.108:0/2980308931' entity='mgr.np0005537446.nfulyg' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch Nov 27 05:19:22 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:22 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:22 localhost nova_compute[284026]: 2025-11-27 10:19:22.318 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:23 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:23 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v783: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s wr, 0 op/s Nov 27 05:19:23 localhost nova_compute[284026]: 2025-11-27 10:19:23.804 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:25 localhost ceph-mgr[290377]: [progress INFO root] Writing back 50 completed events Nov 27 05:19:25 localhost ceph-mon[305520]: mon.np0005537446@2(peon) e17 handle_command mon_command([{prefix=config-key set, key=mgr/progress/completed}] v 0) Nov 27 05:19:25 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v784: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 426 B/s wr, 0 op/s Nov 27 05:19:25 localhost openstack_network_exporter[244641]: ERROR 10:19:25 appctl.go:131: Failed to prepare call to ovsdb-server: no control socket files found for the ovs db server Nov 27 05:19:25 localhost openstack_network_exporter[244641]: ERROR 10:19:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:19:25 localhost openstack_network_exporter[244641]: ERROR 10:19:25 appctl.go:144: Failed to get PID for ovn-northd: no control socket files found for ovn-northd Nov 27 05:19:25 localhost openstack_network_exporter[244641]: ERROR 10:19:25 appctl.go:174: call(dpif-netdev/pmd-perf-show): please specify an existing datapath Nov 27 05:19:25 localhost openstack_network_exporter[244641]: Nov 27 05:19:25 localhost openstack_network_exporter[244641]: ERROR 10:19:25 appctl.go:174: call(dpif-netdev/pmd-rxq-show): please specify an existing datapath Nov 27 05:19:25 localhost openstack_network_exporter[244641]: Nov 27 05:19:26 localhost ceph-mon[305520]: from='mgr.34541 ' entity='mgr.np0005537446.nfulyg' Nov 27 05:19:27 localhost nova_compute[284026]: 2025-11-27 10:19:27.357 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:27 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v785: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 85 B/s wr, 0 op/s Nov 27 05:19:28 localhost ceph-mon[305520]: mon.np0005537446@2(peon).osd e279 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 343932928 full_alloc: 348127232 kv_alloc: 318767104 Nov 27 05:19:28 localhost nova_compute[284026]: 2025-11-27 10:19:28.848 284030 DEBUG ovsdbapp.backend.ovs_idl.vlog [-] [POLLIN] on fd 20 __log_wakeup /usr/lib64/python3.9/site-packages/ovs/poller.py:263#033[00m Nov 27 05:19:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db. Nov 27 05:19:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e. Nov 27 05:19:28 localhost systemd[1]: Started /usr/bin/podman healthcheck run d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132. Nov 27 05:19:28 localhost podman[335411]: 2025-11-27 10:19:28.995856013 +0000 UTC m=+0.081780468 container health_status d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, health_status=healthy, vcs-type=git, url=https://catalog.redhat.com/en/search?searchType=containers, managed_by=edpm_ansible, name=ubi9-minimal, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., release=1755695350, config_id=edpm, maintainer=Red Hat, Inc., architecture=x86_64, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, vendor=Red Hat, Inc., version=9.6, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, container_name=openstack_network_exporter, summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., io.openshift.expose-services=, com.redhat.component=ubi9-minimal-container, io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., distribution-scope=public, build-date=2025-08-20T13:12:41, io.openshift.tags=minimal rhel9, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal) Nov 27 05:19:29 localhost podman[335411]: 2025-11-27 10:19:29.012964743 +0000 UTC m=+0.098889178 container exec_died d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132 (image=quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7, name=openstack_network_exporter, managed_by=edpm_ansible, build-date=2025-08-20T13:12:41, com.redhat.component=ubi9-minimal-container, release=1755695350, config_id=edpm, config_data={'image': 'quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7', 'restart': 'always', 'recreate': True, 'privileged': True, 'ports': ['9105:9105'], 'command': [], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'OPENSTACK_NETWORK_EXPORTER_YAML': '/etc/openstack_network_exporter/openstack_network_exporter.yaml'}, 'healthcheck': {'test': '/openstack/healthcheck openstack-netwo', 'mount': '/var/lib/openstack/healthchecks/openstack_network_exporter'}, 'volumes': ['/var/lib/openstack/config/telemetry/openstack_network_exporter.yaml:/etc/openstack_network_exporter/openstack_network_exporter.yaml:z', '/var/run/openvswitch:/run/openvswitch:rw,z', '/var/lib/openvswitch/ovn:/run/ovn:rw,z', '/proc:/host/proc:ro', '/var/lib/openstack/healthchecks/openstack_network_exporter:/openstack:ro,z']}, url=https://catalog.redhat.com/en/search?searchType=containers, com.redhat.license_terms=https://www.redhat.com/en/about/red-hat-end-user-license-agreements#UBI, io.buildah.version=1.33.7, maintainer=Red Hat, Inc., vcs-type=git, description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., summary=Provides the latest release of the minimal Red Hat Universal Base Image 9., architecture=x86_64, io.openshift.expose-services=, version=9.6, vcs-ref=f4b088292653bbf5ca8188a5e59ffd06a8671d4b, io.openshift.tags=minimal rhel9, container_name=openstack_network_exporter, io.k8s.display-name=Red Hat Universal Base Image 9 Minimal, name=ubi9-minimal, distribution-scope=public, vendor=Red Hat, Inc., io.k8s.description=The Universal Base Image Minimal is a stripped down image that uses microdnf as a package manager. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly.) Nov 27 05:19:29 localhost systemd[1]: d22cb11217d4cc34894a6faec54bed8efc690800281158675e8dd0f4efb22132.service: Deactivated successfully. Nov 27 05:19:29 localhost systemd[1]: tmp-crun.zXz88z.mount: Deactivated successfully. Nov 27 05:19:29 localhost podman[335410]: 2025-11-27 10:19:29.111812228 +0000 UTC m=+0.200886687 container health_status a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, health_status=healthy, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible) Nov 27 05:19:29 localhost podman[335409]: 2025-11-27 10:19:29.158853902 +0000 UTC m=+0.248220019 container health_status 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, health_status=healthy, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, container_name=ceilometer_agent_compute, managed_by=edpm_ansible, tcib_managed=true, maintainer=OpenStack Kubernetes Operator team, org.label-schema.name=CentOS Stream 9 Base Image, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 05:19:29 localhost podman[335409]: 2025-11-27 10:19:29.17070942 +0000 UTC m=+0.260075507 container exec_died 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db (image=quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified, name=ceilometer_agent_compute, org.label-schema.license=GPLv2, maintainer=OpenStack Kubernetes Operator team, container_name=ceilometer_agent_compute, tcib_managed=true, config_id=edpm, io.buildah.version=1.41.3, org.label-schema.build-date=20251125, org.label-schema.name=CentOS Stream 9 Base Image, managed_by=edpm_ansible, org.label-schema.schema-version=1.0, tcib_build_tag=1f5c0439f2433cb462b222a5bb23e629, org.label-schema.vendor=CentOS, config_data={'image': 'quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified', 'user': 'ceilometer', 'restart': 'always', 'command': 'kolla_start', 'security_opt': 'label:type:ceilometer_polling_t', 'net': 'host', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'OS_ENDPOINT_TYPE': 'internal'}, 'healthcheck': {'test': '/openstack/healthcheck compute', 'mount': '/var/lib/openstack/healthchecks/ceilometer_agent_compute'}, 'volumes': ['/var/lib/openstack/config/telemetry:/var/lib/openstack/config/:z', '/var/lib/openstack/config/telemetry/ceilometer-agent-compute.json:/var/lib/kolla/config_files/config.json:z', '/run/libvirt:/run/libvirt:shared,ro', '/etc/hosts:/etc/hosts:ro', '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', '/etc/localtime:/etc/localtime:ro', '/etc/pki/ca-trust/source/anchors:/etc/pki/ca-trust/source/anchors:ro', '/var/lib/openstack/cacerts/telemetry/tls-ca-bundle.pem:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem:ro,z', '/dev/log:/dev/log', '/var/lib/openstack/healthchecks/ceilometer_agent_compute:/openstack:ro,z']}) Nov 27 05:19:29 localhost podman[335410]: 2025-11-27 10:19:29.171934543 +0000 UTC m=+0.261009012 container exec_died a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e (image=quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd, name=podman_exporter, config_id=edpm, container_name=podman_exporter, maintainer=Navid Yaghoobi , managed_by=edpm_ansible, config_data={'image': 'quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd', 'restart': 'always', 'recreate': True, 'user': 'root', 'privileged': True, 'ports': ['9882:9882'], 'net': 'host', 'environment': {'OS_ENDPOINT_TYPE': 'internal', 'CONTAINER_HOST': 'unix:///run/podman/podman.sock'}, 'healthcheck': {'test': '/openstack/healthcheck podman_exporter', 'mount': '/var/lib/openstack/healthchecks/podman_exporter'}, 'volumes': ['/run/podman/podman.sock:/run/podman/podman.sock:rw,z', '/var/lib/openstack/healthchecks/podman_exporter:/openstack:ro,z']}) Nov 27 05:19:29 localhost systemd[1]: a77ba3991cd3e37180b8bb73bbd96aebaae6573a5a2ff462afb67287ad13346e.service: Deactivated successfully. Nov 27 05:19:29 localhost systemd[1]: 9cda1aaf3051196ae8158cc0486e267a9884710fa5ca47beb57ec117174c74db.service: Deactivated successfully. Nov 27 05:19:29 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v786: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 0 B/s wr, 0 op/s Nov 27 05:19:31 localhost ceph-mgr[290377]: log_channel(cluster) log [DBG] : pgmap v787: 177 pgs: 177 active+clean; 227 MiB data, 1.3 GiB used, 41 GiB / 42 GiB avail; 0 B/s wr, 0 op/s Nov 27 05:19:31 localhost sshd[335469]: main: sshd: ssh-rsa algorithm is disabled Nov 27 05:19:32 localhost systemd-logind[761]: New session 77 of user zuul. Nov 27 05:19:32 localhost systemd[1]: Started Session 77 of User zuul.